HACKER Q&A
📣 Rizu

Recommendation for a SWE looking to get up to speed with latest on AI


I am looking to get up to speed with the latest things happening in AI, I use ChatGPT almost everyday and i last used the open AI api for 3.5 last year. I am looking for a tech blogs like HN to keep updated on things AI, I came across https://simonwillison.net/ but it appears fragmented


  👤 tikkun Accepted Answer ✓
It sounds like you want more broad stuff, not necessarily learning how to train models. More like learning to use them and how they work.

https://news.ycombinator.com/item?id=36195527 and

Hacker's Guide to LLMs by Jeremy from Fast.ai - https://www.youtube.com/watch?v=jkrNMKz9pWU

State of GPT by Karpathy - https://www.youtube.com/watch?v=bZQun8Y4L2A

LLMs by 3b1b - https://www.youtube.com/watch?v=LPZh9BOjkQs

Visualizing transformers by 3b1b - https://www.youtube.com/watch?v=KJtZARuO3JY

How ChatGPT trained - https://www.youtube.com/watch?v=VPRSBzXzavo

AI in a nutshell - https://www.youtube.com/watch?v=2IK3DFHRFfw

How Carlini uses LLMs - https://nicholas.carlini.com/writing/2024/how-i-use-ai.html

For staying updated:

X/Twitter & Bluesky. Go and follow people that work at OpenAI, Anthropic, Google DeepMind, and xAI.

Podcasts: No Priors, Generally Intelligent, Dwarkesh Patel, Sequoia's "Training Data"


👤 pdevine
The poster's looking for articles, so this recommendation's a bit off the mark. I learned more from participating in a few Kaggle competitions (https://www.kaggle.com/competitions) than I did from reading about AI. Many folks in the community shared their homework, and by learning how to follow their explanations I developed a much more intuitive understanding of the technology. The first competition had a steep learning curve. I felt it was worth it. The application of having a specific goal and the provided datasets made the problem space more tractable.


👤 Maro
I don't think it's a good idea to kepp up to date at a daily/weekly cadence, unless you somehow directly get paid for it. It's like checking stocks daily, it doesn't lead to good investment decisions.

It's better to do it more batchy, like once every 6-12 months or so.


👤 explaingarlic
So I'm currently using "OpenCV University"'s playlist on YouTube to get myself up to speed with computer vision, and this has lead into a spiraling staircase down into the depths of CNNs.

Started off here: https://www.youtube.com/watch?v=hZWgEPOVnuM&list=PL6e-Bu0cqf...

Ended up here: https://www.youtube.com/watch?v=_5XYLA2HLmo&list=PL6e-Bu0cqf...

And after that, I've had some recent projects that I love to mess around with such as a better license plate detection API than what currently exists for U.K. plates, and once I completed those two courses I had a good enough baseline to work from where I'd encounter a repository and google around if I needed to learn something new.

Short, simple, not painful etc. and I don't have the advanced mathematical background (nor the background within the American mathematical notation) that I'd need to digest the MIT course set, so this learning path has been the best for me. I'm no expert whatsoever, though.


👤 zackmorris
LLMs and neural nets from first principles:

https://arxiv.org/pdf/2404.17625 (pdf)

https://news.ycombinator.com/item?id=40408880 (llama3 implementation)

https://news.ycombinator.com/item?id=40417568 (my comment on llama3 with breadcrumbs)

Admittedly, I'm way behind on how this translates to software on the newest video cards. Part of that is that I don't like the emphasis on GPUs. We're only seeing the SIMD side of deep learning with large matrices and tensors. But there are at least a dozen machine learning approaches that are being neglected, mainly genetic algorithms. Which means that we're perhaps focused too much on implementations and not on core algorithms. It would be like trying to study physics without change of coordinates, Lorentz transformations or calculus. Lots of trees but no forest.

To get back to rapid application development in machine learning, I'd like to see a 1000+ core, 1+ GHz CPU with 16+ GBs of core-local ram for under $1000 so that we don't have to manually transpile our algorithms to GPU code. That should have arrived around 2010 but the mobile bubble derailed desktop computing. Today it should be more like 10,000+ cores for that price at current transistor counts, increasing by a factor of about 100 each decade by what's left of Moore's law.

We also need better languages. Something like a hybrid of Erlang and Go with always-on auto-parallelization to run our human-readable but embarrassingly parallel code.

Short of that, there might be an opportunity to write a transpiler that converts C-style imperative or functional code to existing GPU code like CUDA (MIMD -> SIMD). Julia is the only language I know of even trying to do this.

Those are the areas where real work is needed to democratize AI, that SWEs like us may never be able to work on while we're too busy making rent. And the big players like OpenAI and Nvidia have no incentive to pursue them and disrupt themselves.

Maybe someone can find a challenging profit where I only see disillusionment, and finally deliver UBI or at least stuff like 3D printed robots that can deliver the resources we need outside of a rigged economy.


👤 senko
I follow these:

* Matt Berman on X / YT

* AI-summarized AI news digest: https://buttondown.com/ainews by swyx

* https://codingwithintelligence.com/about by Rick Lamers

Then I manually follow up to learn more about specific topic/news I'm interested in.


👤 fallinditch
New short course on FreeCodeCamp YouTube channel looks good -

Ollama Course – Build AI Apps Locally https://youtu.be/GWB9ApTPTv4?feature=shared

As an aside, does anyone have any ideas about this: there should be an app like an 'auto-RAG' that scrapes RSS feeds and URLs, in addition to ingesting docs, text and content in the normal RAG way. Then you could build AI chat-enabled knowledge resources around specific subjects. Autogenerated summaries and dashboards would provide useful overviews.

Perhaps this already exists?


👤 adroitboss
The best place for the latest information isn't tech blogs in my opinion. It's the stable diffusion and local llama subreddits. If you are looking to learn about everything on a fundamental level you need to check out Andrej Karpathy on YouTube. There other some other notable mentions in other people's comments.

👤 bmitc
Are you wanting to get into LLMs in particular or something else? I am a software engineer also trying to make headways into so-called "AI", but I have little interest in LLMs. For one, it's suffering from a major hype bubble right now. The second reason is that because of reason one, it has a huge amount of attention from people who study and work on this every day. It's not something I have the time commitment for to compete with that. Lastly, as mentioned, I have no interest in it and my understanding of them leads me to believe they have few interesting applications besides generating a huge amount of noise in society and dumping heat. The Internet, like blogs, articles, and even YouTube, are already being overrun by LLM-generated material that is effectively worthless. I'm not sure of the net positive for LLMs.

For me personally, I prefer to work backwards and then forwards. What I mean by that is that I want to understand the basics and fundamentals first. So, I'm, slowly, trying to bone up on my statistics, probability, and information theory and have targeted machine learning books that also take a fundamental approach. There's no end to books in this realm for neural networks, machine learning, etc., so it's hard to recommend beyond what I've just picked, and I'm just getting started anyway.

If you can get your employer to pay for it, MIT xPRO has courses on machine learning (https://xpro.mit.edu/programs/program-v1:xPRO+MLx/ and https://xpro.mit.edu/courses/course-v1:xPRO+GenAI/). These will likely give a pretty up to date overview of the technologies.


👤 canyon289
I work on frontier models directly and I I wrote a guidebook for folks like you. https://ravinkumar.com/GenAiGuidebook/book_intro.html

It has a mix of concepts and hands on code, and lots of links to the best places to learn more. I'm keeping it up to date as well, about to merge a guide on building applications, which is what it sounds like you want.

Here's my Google scholar if you want credentials https://scholar.google.com/citations?user=Oq99ddEAAAAJ&hl=en...


👤 jayalammar
We actually just wrote a book with your profile in mind -- especially if by "AI" you're especially interested in LLMs and if you're a visual learner. It's called Hands-On Large Language Models and it contains 300 original figures explaining the main couple hundred intuitions and applications for these models. You can also read it online on the O'Reilly platform. I find that after acquiring the main intuitions, people find it much easier to move on to code implementations or papers.

👤 AlphaWeaver
As I was building up my understanding/intuition for the internals of transformers + attention, I found 3Blue1Brown's series of videos (specifically on attention) to be super helpful.

👤 ohgrrr
Using web wrappers around AI is being an AI end user

If you want to be an AI engineer study this:

https://github.com/karpathy/llm.c

And build around llama.cpp

Ollama is like cpanel for models. It’s not going to familiarize you with lower level implementation which is just as important as knowing the math.

That was my approach. Being aware of the internals not just the equivalent of “git pull model” got me a job, without a CS degree and a long career in software. Ymmv


👤 petesergeant
Read through this making flashcards as you to: https://eugeneyan.com/writing/llm-patterns/

Then spin up a RAG-enhanced chatbot using pgvector on your favourite subject, and keep improving it when you learn about cool techniques


👤 simonw
My blog is very high volume so yeah, it can be difficult to know where to look on it.

I use tags a lot - these ones might be more useful for you:

https://simonwillison.net/tags/prompt-engineering/ - collects notes on prompting techniques

https://simonwillison.net/tags/llms/ - everything relating to LLMs

https://simonwillison.net/tags/openai/ and https://simonwillison.net/tags/anthropic/ and https://simonwillison.net/tags/gemini/ and https://simonwillison.net/tags/llama/ and https://simonwillison.net/tags/mistral/ - I have tags for each of the major model families and vendors

Every six months or so I write something (often derived from a conference talk) that's more of a "catch up with the latest developments" post - a few of those:

- Stuff we figured out about AI in 2023 - https://simonwillison.net/2023/Dec/31/ai-in-2023/ - I will probably do one of those for 2024 next month

- Imitation Intelligence, my keynote for PyCon US 2024 - https://simonwillison.net/2024/Jul/14/pycon/ from July this year


👤 BillFranklin
I read about 30 LLM papers a couple months ago dated from 2018-2024. Mostly folks are publishing on the “how do we prompt better” problem, and you can kind of get the gist in about a day by reading a few blogs (RAG, fine tuning, tool use, etc). There is also more progress being made for model capabilities, like multi modality, and each company seems to be pushing in only slightly different directions, but essentially they are still black boxes.

It depends what you are looking for honestly “the latest things happening” is pretty vague. I’d say the place to look is probably just the blogs of OpenAI/Anthropic/Genini, since they are the only teams with inside information and novel findings to report. Everyone else is just using the tools we are given.


👤 iamwil
Lots of people can get impressive demos up and running, but if you want to run AI products in production, you're going to have to do system evals. System evals make sure your product is doing what it says on the box with unquantifiable qualities.

We wrote a zine on system evals without jargon: https://forestfriends.tech

Eugene Yan has written extensively on it https://eugeneyan.com/writing/evals/

Hamel has as well. https://hamel.dev/blog/posts/evals/


👤 zellyn
Simon's blog is fragmented because it's, well, a blog. It would be hard to find a better source to "keep updated on things AI" though. He does do longer summary articles sometimes, but mostly he's keeping up with things in real time. The search and tagging systems on his blog work well, too. I suggest you stick his RSS feed in your feed reader, and follow along that way.

Swyx also has a lot of stuff keeping up to date at https://www.latent.space/, including the Latent Space podcast, although tbh I haven't listened to more than one or two episodes.


👤 danofsteel32
I recently wrote a post for a coworker who asked the exact same question.

https://dandavis.dev/llm-knowledge-dump.html



👤 handzhiev
For news-like content I follow accounts on X: @kimmonismus @apples_jimmy and the accounts of Antropic, Mistal, Gemini / DeepMind and OpenAI. I think everyone who is really interested in the hot AI developments must also follow what comes from China. I follow https://chinai.substack.com/ but I am open to hear about other Chinese resources.

👤 notslow
Machine Learning Mastery (https://machinelearningmastery.com) provides code examples for many of the popular models. For me, seeing and writing code has been helpful in understanding how things work and makes it easier to put new developments in context.

👤 ketanmaheshwari

👤 bingemaker
Being a coder, I find these resources extremely useful:

Github blog: https://github.blog/ai-and-ml/ Cursor blog: https://www.cursor.com/blog


👤 Workaccount2
The localllama subreddit, although focused mostly on open source locally run models, still has ample discussion of SOTA models too.

https://old.reddit.com/r/LocalLLaMA/


👤 mavelikara
I found video lectures of “Advanced NLP” course by Mohit Iyer very useful to get me started: https://people.cs.umass.edu/~miyyer/cs685/

👤 jumping_frog

👤 Exorust
I've been making a breakdown on the topics present for LLMs https://charoori.notion.site/topicwise-breakdown

👤 aaronrobinson
What a goldmine of recommendations. I like Sam Witterveen’s YouTube stuff for keeping up to speed https://m.youtube.com/@samwitteveenai

👤 goosethe
https://playground.tensorflow.org/ this is a classic which, imo, breaks it down to the simplest visuals.

👤 ode
ThursdAI - all the best AI news from the last week - https://thursdai.news/

They also have a weekly podcast.


👤 brcmthrowaway
Who else bookmarked this Ask HN thread never to revisit?

👤 eachro
Reproduce nanogpt.

Then find a small dataset and see if you can start getting close to some of the reported benchmark numbers with similar architectures.


👤 cranberryturkey
checkout ollama. it lets you run open models on your own hardware. it also provides an easy to use rest api similar to openai's

👤 febin
Build a tool on top of the LLM layer for a specific use case. That'll get you up to speed. You haven't missed much.

👤 toddwprice
Subscribe to The Neuron newsletter

👤 gargigupta97
Unwind AI would be helpful. They publish daily newsletters on AI as well as tutorials on building apps with step-by-step walkthrough. Super focused on developers. https://www.theunwindai.com/

👤 mindcrime
Lots of good suggestions here already. I'd start by adding one quick note though. "AI" is more than just LLM's. Sure, the "current, trendy, fashionable" thing is all LLM's, but the field as a whole is still much larger. I'd encourage you to not myopically focus on LLM's to exclusion. Depending on your existing background knowledge, there's a lot to be said for going out and getting a copy of Artificial Intelligence: A Modern Approach and reading through it. Likewise for something like Hands-On Machine Learning with Scikit-Learn, Keras, and Tensorflow.

Beyond that: there are some decent sub-reddits for keeping up with AI happenings, a lot of good Youtube channels (although a lot of the ones that talk about the "current, trendy" AI stuff tend to be a bit tabloid'ish), and even a couple of Facebook groups. You can also find good signal by choosing the right people to follow on Twitter/LinkedIn/Mastodon/Bluesky/etc.

https://www.reddit.com/r/artificial/

https://reddit.com/r/machineLearning/

https://www.reddit.com/r/LLM/

https://www.reddit.com/r/agi

https://www.reddit.com/r/ollama/

https://www.youtube.com/@matthew_berman

https://www.youtube.com/@TheAiGrid

https://www.youtube.com/@WesRoth

https://www.youtube.com/@DaveShap

https://www.youtube.com/c/MachineLearningStreetTalk

https://www.youtube.com/@twimlai

https://www.youtube.com/@YannicKilcher

And you can always go straight to "the source" and follow pre-prints showing up in arXiv.

https://arxiv.org/corr

For tools to make it easier to track new releases, arXiv supports subscriptions to daily digest emails, and also has RSS feeds.

https://info.arxiv.org/help/subscribe.html

https://info.arxiv.org/help/rss.html

There are also some bots in the Fediverse that push out links to new arXiv papers.


👤 sveinek
I found the YouTube channel Data Centric very informative and useful.

👤 barrenko
Get on Twitter (well, X) as that's where the the cutting edge is.

👤 aanet
Excellent thread! Love the responses.

Is there a way to SAVE THIS THREAD on HN ? 'Cos I'd love that.

Thx


👤 maeil
Subscribe to the TLDR newsletter (https://tldr.tech/) and read the AI-related articles it links. No personal affiliation, just a satisfied reader.

👤 JSDevOps
First thing you need to do is change your LinkedIn to “AI evangelist” then go to your boss and say I want triple the pay. Then let the chips fall where they may. Oh also rename all your GitHub or personal projects to have AI in the name. You don’t actually have to do much else.

👤 not_your_vase
Unpopular opinion: if you can't use Google nor ChatGPT to get an answer to this question, I have bad news for you.