As you’ll have heard, DeepSeek-R1 is making waves. It’s everywhere in the AI newsfeed, hailed as the primary open-source reasoning mannequin of its sort.
The thrill? Effectively-deserved.
The mannequin? Highly effective.
DeepSeek-R1 represents the present frontier in reasoning fashions, pushing the boundaries of what open-source AI can obtain. However right here’s the half you gained’t see within the headlines: working with it isn’t precisely simple.
Prototyping might be clunky. Deploying to manufacturing? Even trickier.
That’s the place DataRobot is available in. We make it simpler to develop with and deploy DeepSeek-R1, so you possibly can spend much less time wrestling with complexity and extra time constructing actual, enterprise-ready options.
Prototyping DeepSeek-R1 and bringing functions into manufacturing are crucial to harnessing its full potential and delivering higher-quality generative AI experiences.
So, what precisely makes DeepSeek-R1 so compelling — and why is it sparking all this consideration? Let’s take a more in-depth look to see if all of the hype is justified.
May this be the mannequin that outperforms OpenAI’s newest and biggest?
Past the hype: Why DeepSeek-R1 is value your consideration
DeepSeek-R1 isn’t simply one other generative AI mannequin. It’s arguably the primary open-source “reasoning” mannequin — a generative textual content mannequin particularly strengthened to generate textual content that approximates its reasoning and decision-making processes.
For AI practitioners, that opens up new potentialities for functions that require structured, logic-driven outputs.
What additionally stands out is its effectivity. Coaching DeepSeek-R1 reportedly price a fraction of what it took to develop fashions like GPT-4o, due to reinforcement studying methods printed by DeepSeek AI. And since it’s totally open-source, it provides higher flexibility whereas permitting you to keep up management over your knowledge.
In fact, working with an open-source mannequin like DeepSeek-R1 comes with its personal set of challenges, from integration hurdles to efficiency variability. However understanding its potential is step one to creating it work successfully in real-world functions and delivering extra related and significant experiences to finish customers.
Utilizing DeepSeek-R1 in DataRobot
In fact, potential doesn’t all the time equal straightforward. That’s the place DataRobot is available in.
With DataRobot, you possibly can host DeepSeek-R1 utilizing NVIDIA GPUs for high-performance inference or entry it by way of serverless predictions for quick, versatile prototyping, experimentation, and deployment.
Regardless of the place DeepSeek-R1 is hosted, you possibly can combine it seamlessly into your workflows.
In follow, this implies you possibly can:
- Evaluate efficiency throughout fashions with out the effort, utilizing built-in benchmarking instruments to see how DeepSeek-R1 stacks up towards others.
- Deploy DeepSeek-R1 in manufacturing with confidence, supported by enterprise-grade safety, observability, and governance options.
- Construct AI functions that ship related, dependable outcomes, with out getting slowed down by infrastructure complexity.
LLMs like DeepSeek-R1 are hardly ever utilized in isolation. In real-world manufacturing functions, they operate as a part of refined workflows quite than standalone fashions. With this in thoughts, we evaluated DeepSeek-R1 inside a number of retrieval-augmented technology (RAG) pipelines over the well-known FinanceBench dataset and in contrast its efficiency to GPT-4o mini.
So how does DeepSeek-R1 stack up in real-world AI workflows? Right here’s what we discovered:
- Response time: Latency was notably decrease for GPT-4o mini. The eightieth percentile response time for the quickest pipelines was 5 seconds for GPT-4o mini and 21 seconds for DeepSeek-R1.
- Accuracy: The perfect generative AI pipeline utilizing DeepSeek-R1 because the synthesizer LLM achieved 47% accuracy, outperforming the very best pipeline utilizing GPT-4o mini (43% accuracy).
- Price: Whereas DeepSeek-R1 delivered greater accuracy, its price per name was considerably greater—about $1.73 per request in comparison with $0.03 for GPT-4o mini. Internet hosting selections impression these prices considerably.
Whereas DeepSeek-R1 demonstrates spectacular accuracy, its greater prices and slower response instances could make GPT-4o mini the extra environment friendly selection for a lot of functions, particularly when price and latency are crucial.
This evaluation highlights the significance of evaluating fashions not simply in isolation however inside end-to-end AI workflows.
Uncooked efficiency metrics alone don’t inform the total story. Evaluating fashions inside refined agentic and non-agentic RAG pipelines provides a clearer image of their real-world viability.
Utilizing DeepSeek-R1’s reasoning in brokers
DeepSeek-R1’s power isn’t simply in producing responses — it’s in the way it causes by way of advanced eventualities. This makes it significantly beneficial for agent-based methods that have to deal with dynamic, multi-layered use circumstances.
For enterprises, this reasoning functionality goes past merely answering questions. It might:
- Current a spread of choices quite than a single “greatest” response, serving to customers discover completely different outcomes.
- Proactively collect info forward of consumer interactions, enabling extra responsive, context-aware experiences.
Right here’s an instance:
When requested concerning the results of a sudden drop in atmospheric strain, DeepSeek-R1 doesn’t simply ship a textbook reply. It identifies a number of methods the query could possibly be interpreted — contemplating impacts on wildlife, aviation, and inhabitants well being. It even notes much less apparent penalties, just like the potential for out of doors occasion cancellations resulting from storms.
In an agent-based system, this sort of reasoning might be utilized to real-world eventualities, akin to proactively checking for flight delays or upcoming occasions that is likely to be disrupted by climate modifications.
Curiously, when the identical query was posed to different main LLMs, together with Gemini and GPT-4o, none flagged occasion cancellations as a possible danger.
DeepSeek-R1 stands out in agent-driven functions for its skill to anticipate, not simply react.

Evaluate DeepSeek-R1 to GPT 4o-mini: What the info tells us
Too typically, AI practitioners rely solely on an LLM’s solutions to find out if it’s prepared for deployment. If the responses sound convincing, it’s straightforward to imagine the mannequin is production-ready. However with out deeper analysis, that confidence might be deceptive, as fashions that carry out properly in testing typically wrestle in real-world functions.
That’s why combining professional assessment with quantitative assessments is crucial. It’s not nearly what the mannequin says, however the way it will get there—and whether or not that reasoning holds up below scrutiny.
As an instance this, we ran a fast analysis utilizing the Google BoolQ studying comprehension dataset. This dataset presents brief passages adopted by sure/no questions to check a mannequin’s comprehension.
For GPT-4o-mini, we used the next system immediate:
Attempt to reply with a transparent YES or NO. You may additionally say TRUE or FALSE however be clear in your response.
Along with your reply, embody your reasoning behind this reply. Enclose this reasoning with the tag
For instance, if the consumer asks “What shade is a can of coke” you’ll say:
Reply: Purple
Right here’s what we discovered:
- Proper: DeepSeek-R1’s output.
- On the far left: GPT-4o-mini answering with a easy Sure/No.
- Heart: GPT-4o-mini with reasoning included.

We used DataRobot’s integration with LlamaIndex’s correctness evaluator to grade the responses. Curiously, DeepSeek-R1 scored the bottom on this analysis.

What stood out was how including “reasoning” brought on correctness scores to drop throughout the board.
This highlights an necessary takeaway: whereas DeepSeek-R1 performs properly in some benchmarks, it could not all the time be the very best match for each use case. That’s why it’s crucial to check fashions side-by-side to search out the appropriate software for the job.
Internet hosting DeepSeek-R1 in DataRobot: A step-by-step information
Getting DeepSeek-R1 up and operating doesn’t need to be sophisticated. Whether or not you’re working with one of many base fashions (over 600 billion parameters) or a distilled model fine-tuned on smaller fashions like LLaMA-70B or LLaMA-8B, the method is simple. You possibly can host any of those variants on DataRobot with only a few setup steps.
1. Go to the Mannequin Workshop:
- Navigate to the “Registry” and choose the “Mannequin Workshop” tab.

2. Add a brand new mannequin:
- Title your mannequin and select “[GenAI] vLLM Inference Server” below the atmosphere settings.
- Click on “+ Add Mannequin” to open the Customized Mannequin Workshop.

3. Arrange your mannequin metadata:
- Click on “Create” so as to add a model-metadata.yaml file.

4. Edit the metadata file:
- Save the file, and “Runtime Parameters” will seem.
- Paste the required values from our GitHub template, which incorporates all of the parameters wanted to launch the mannequin from Hugging Face.

5. Configure mannequin particulars:
- Choose your Hugging Face token from the DataRobot Credential Retailer.
- Below “mannequin,” enter the variant you’re utilizing. For instance: deepseek-ai/DeepSeek-R1-Distill-Llama-8B.
6. Launch and deploy:
- As soon as saved, your DeepSeek-R1 mannequin might be operating.
- From right here, you possibly can take a look at the mannequin, deploy it to an endpoint, or combine it into playgrounds and functions.
From DeepSeek-R1 to enterprise-ready AI
Accessing cutting-edge generative AI instruments is simply the beginning. The actual problem is evaluating which fashions suit your particular use case—and safely bringing them into manufacturing to ship actual worth to your finish customers.
DeepSeek-R1 is only one instance of what’s achievable when you might have the flexibleness to work throughout fashions, evaluate their efficiency, and deploy them with confidence.
The identical instruments and processes that simplify working with DeepSeek may help you get essentially the most out of different fashions and energy AI functions that ship actual impression.
See how DeepSeek-R1 compares to different AI fashions and deploy it in manufacturing with a free trial.
In regards to the writer

Nathaniel Daly is a Senior Product Supervisor at DataRobot specializing in AutoML and time sequence merchandise. He’s centered on bringing advances in knowledge science to customers such that they will leverage this worth to resolve actual world enterprise issues. He holds a level in Arithmetic from College of California, Berkeley.

Luke Shulman, Technical Discipline Director: Luke has over 15 years of expertise in knowledge analytics and knowledge science. Previous to becoming a member of DataRobot, Luke led implementations and was a director on the product administration workforce at Arcadia.io, the main healthcare SaaS analytics platform. He continued that function at N1Health. At DataRobot, Luke leads integrations throughout the AI/ML ecosystem. He’s additionally an energetic contributor to the DrX extensions to the DataRobot API consumer and MLFlow integration. An avid champion of information science, Luke has additionally contributed to initiatives throughout the info science ecosystem together with Bokeh, Altair, and Zebras.js.