Within the Writer Highlight sequence, TDS Editors chat with members of our group about their profession path in information science and AI, their writing, and their sources of inspiration. Right this moment, we’re thrilled to share our dialog with Vyacheslav Efimov.
Vyacheslav is a Senior Machine Studying Engineer specialising in NLP and Pc Imaginative and prescient. One in every of his passions is making instructional content material: Vyacheslav has printed over 60 articles on In direction of Information Science, explaining complicated ideas in easy phrases, offering wealthy visualizations.
You wrote many newbie and explanatory articles on TDS. Has educating the basics modified the way you design or debug actual methods at work?
I discover the correlation between the extra I educate one thing, the higher I perceive it. In actual life, after I write a brand new article, I try to dive into small particulars whereas protecting the reason easy for my readers. Emphasizing data on this manner helps me higher perceive the workflow of algorithms.
In that sense, each time I encounter an error in one of many algorithms used at work on which I wrote an article previously, there’s a increased likelihood that I’ll quickly work out the answer to the issue on my own. From one other perspective, after I write an article on an unfamiliar subject and discover it myself, it will increase my confidence after I apply that individual algorithm at work, as I already know its software scope, benefits, disadvantages, and particular particulars or constraints.
This fashion, I can provide you with unique options that aren’t apparent to others and again up my option to different teammates, managers, or stakeholders. That data is valuable to me.
With so many new fashions popping up every single day, it’s simple to really feel fully swamped. How do you determine what’s price a ‘deep dive’ and what you simply ‘get the gist of’? Has your technique for managing this modified in any respect just lately?
Right this moment, we certainly have an abundance of fashions and instruments that seem every single day. It’s simple to really feel misplaced whenever you’re not sure about what to pursue subsequent.
With restricted time, I often delve deeper into matters which may be relevant at work or in my private initiatives. This offers me extra confidence when I’ve to current or clarify my outcomes.
Companies often need to obtain working outcomes as quickly as potential. That is additionally one of many the explanation why, in my articles, I focus extra on theoretical ideas, as I can’t dedicate my time at work to going into theoretical depth.
This fashion, I’ve an environment friendly mix of sensible expertise at work and theoretical insights in my weblog. Each of those parts are necessary for expert information scientists.
You’ve competed in AI hackathons. What did you be taught from having such tight deadlines? Did it pressure you to get higher at scoping initiatives or deciding on a mannequin? And do you end up utilizing any of these ‘hackathon classes’ whenever you’re sketching out a brand new concept from scratch?
Hackathons sometimes final between a number of hours and two days. That could be a very small time frame to develop a totally purposeful product. Nevertheless, on the similar time, it pushed me loads previously to raised prioritize the options on which I ought to focus. Generally, time administration is a invaluable ability to have. When you’ve a number of potential options to handle your drawback, you need to select the one that most closely fits the enterprise wants whereas additionally respecting time constraints.
What can also be nice is that after each hackathon, you may consider your self by way of the time it took you to implement sure options. For instance, let’s say that it was the primary time you needed to develop a RAG pipeline, which took you round 4 hours to implement. The following time you face an identical drawback at work or a hackathon, you should have a greater estimate upfront of how a lot time it might take in case you determine to make use of the identical technique. In that sense, the hackathon expertise lets you higher outline closing dates for the strategies you need to implement in initiatives.
For me, the largest lesson from the hackathon was not specializing in perfection when creating the MVP. Whereas an MVP is necessary, it’s also essential to current your product attractively to purchasers or buyers, clarify its enterprise worth, the issue it solves, and why it’s higher than present options in the marketplace. On this regard, hackathons educate you to provide you with higher concepts that clear up actual issues whereas additionally transport the MVP shortly, containing probably the most important options.
For readers serious about their profession path: your “Roadmap to Turning into a Information Scientist” sequence spans fundamentals by superior ML. When you have been rewriting it right this moment, what matters would get promoted, demoted, or lower completely, and why?
I wrote this text sequence a 12 months in the past. For me, all of the ideas and matters I listed are updated for aspiring information scientists. All math, pc science, and machine studying matters I current there are a necessary basis for any machine studying engineer.
As we’re now in late 2025, I’d additionally add a requirement to have no less than minimal expertise with immediate engineering and to be conversant in some AI-generative instruments, reminiscent of GitHub Copilot, Gemini CLI, and Cursor, which might permit for elevated work effectivity.
As a notice, in comparison with earlier years, IT firms have increased necessities and expectations for junior engineers getting into the information science discipline. It is sensible, as trendy AI instruments can carry out junior-level duties very effectively, and plenty of firms desire to depend on them now moderately than on entry-level engineers, as they don’t should pay salaries whereas in each circumstances they obtain the identical end result.
That’s the reason, if a machine studying engineer possesses the sturdy basic expertise I described in that sequence of articles, it is going to be a lot simpler for them to dive autonomously into extra complicated matters.
Your background blends software program engineering and ML. How does that basis form the way in which you write?
Having sturdy software program engineering expertise is likely one of the finest benefits you may have as a Information Scientist:
- It makes you understand the significance of well-structured software program documentation and creating reproducible ML pipelines.
- You perceive higher how you can make your code clear and readable for others.
- You perceive algorithmic constraints and which information construction to decide on for a specific process, primarily based on system wants.
- You possibly can extra simply collaborate with backend and DevOps engineers on integrating your code modules.
- You do not want to depend on others to make SQL queries to retrieve needed information from the database.
The checklist can go on and on…
Talking of my articles, I don’t have many who current loads of code. Nevertheless, each time I do, I try to make it readable and comprehensible to others. I all the time put myself within the footwear of others and ask myself how my article textual content or code examples can be simple to understand or reproduce if I have been in others’ footwear. That is the place the software program engineering expertise makes this realization extra important for me, and I comply with the perfect established practices to ship my last product.
your portfolio and GitHub, you’ve blended software program engineering fundamentals with ML from the beginning. What’s one engineering behavior you want extra aspiring information scientists adopted early?
Many engineers, particularly juniors, are likely to underestimate the significance of making good documentation and reproducible pipelines. This additionally occurred to me previously, after I was extra targeted on growing strong fashions or conducting analysis.
Because it turned out, after I needed to change contexts after which a number of weeks later to return to work on the earlier challenge, I used to be then spending loads of time determining how you can run my outdated code in a messy Jupyter Pocket book or set up needed libraries once more, the place I might have simply spent somewhat extra time previously by growing a well-documented README.md explaining all of the required steps to execute pipelines from zero.
As a result of it was almost unimaginable to rerun my pipelines from scratch, I used to be additionally unable to conduct experiments utilizing different entry parameters, which made the scenario much more irritating.
It was a painful expertise for me, but one of the invaluable classes I’ve discovered. So if I needed to give a bit of recommendation to an aspiring information scientist on one explicit behavior, it might be this:
“At all times make your machine studying pipelines reusable and well-documented”.
Over the previous 12 months, has AI meaningfully modified how you’re employed daily as an ML Engineer? What bought simpler, what bought more durable, and what stayed the identical?
Lately, we’ve got noticed a big rise in highly effective AI engineering instruments:
- LLMs, which might reply to nearly any query, give recommendation, or discover bugs in software program
- Cursor, Lovable, and Bolt are performing as AI-powered IDEs for builders
- AI brokers can full multi-step duties
As a machine studying engineer, it’s important for me to repeatedly adapt to those instruments to make use of them effectively.
What turned simpler
Ranging from 2025, I can observe the next optimistic impression on my work:
- For me, it turned simpler to quickly check concepts or prototypes. For instance, there have been instances at work after I was given pc imaginative and prescient issues that fell outdoors my space of data. On this manner, I might ask ChatGPT to suggest a number of concepts to resolve them. There have been even instances when ChatGPT generated code for me, and I attempted to execute it with out understanding the way it labored inside.
Then I had two potential circumstances:- If the code ran efficiently and solved the preliminary drawback, then I attempted to go deeper contained in the OpenCV documentation to know the way it in the end works.
- If the code didn’t clear up my drawback, I’d both ignore it, report the error to ChatGPT, or try to search out the answer myself.
As you may see, I used to be in a position to quickly check an answer that might work and save me hours of analysis with none threat.
- One other glorious use case for me was inserting error messages instantly into ChatGPT as a substitute of trying to find an answer on the Web. It labored effectively more often than not, however generally it was affected by errors associated to library installations, system errors, and the deployment of pipelines on the Cloud, amongst different points.
- Lastly, I’m an enormous fan of AI hackathons! Having instruments that may generate each the frontend and backend of your system makes an enormous distinction for me, as I can now quickly create prototypes and check my MVP in a number of hours. What I develop now throughout one-day hackathons might require a complete week of labor.
What turned more durable / dangerous
- When writing code with AI, there’s a increased risk of delicate information leaks. Think about you’ve a file or code fragment containing important credentials that you simply by chance feed into an AI mannequin. Then a third-party software will know your delicate credentials. It may possibly occur, particularly in case you use a software like Cursor and retailer your credentials in one other file moderately than .env. As a consequence, it’s all the time essential to be very cautious.
- One other threat will not be correctly testing the AI-generated code and never figuring out how you can make a rollback. An AI software can introduce invisible errors within the code, significantly when it’s used to switch or refactor present code. To make sure that AI-generated code doesn’t degrade, it’s essential to totally evaluation the generated code components, check them, and save modifications in a manner that lets you all the time rollback to a earlier, appropriate model if needed.
- When relying too closely on generative AI instruments, there’s a threat that the code will change into unreadable, include excessively lengthy capabilities, exhibit repetition, or stop to perform accurately. That’s the reason it’s important to know that AI instruments work extra successfully on prototyping than on sustaining high-quality manufacturing code.
What remained the identical
What stays fixed for me is the significance of understanding the inner workflow of the algorithms I exploit, sustaining sturdy pc science foundations, and writing high-quality code, amongst different key expertise. In different phrases, the essential ideas of software program improvement will all the time be essential to effectively use AI instruments.
In that sense, I like evaluating a set of obtainable AI instruments to an alternative choice to a junior developer in my group, to whom I can delegate much less important duties. I can ask it no matter I would like, however I can’t be 100% positive it is going to do my duties accurately, and that is the place the significance of getting sturdy basic experience comes into play.
To be taught extra about Vyacheslav‘s work and keep up-to-date together with his newest articles, you may comply with him on TDS or LinkedIn.

