Interesting Content in AI, Software, Business, and Tech- 5/31/2023

Don’t miss the most important ideas in Machine Learning, AI, Data Science, Business of Tech, and More

Devansh
6 min readJun 4, 2023

A lot of people reach out to me for reading recommendations. I figured I’d start sharing whatever AI Papers/Publications, interesting books, videos, etc I came across each week. Some will be technical, others not really. I will add whatever content I found really informative (and I remembered through the week). These won’t always be the most recent publications- just the ones I’m paying attention to this week. Without further ado, here are interesting readings/viewings for 5/31/2023. If you missed last week’s readings, you can find it here.

Join 32K+ people and get insights on the most important ideas in AI straight to your inbox through my free newsletter- AI Made Simple

AI Papers/Writeups

1) The Larger They Are, the Harder They Fail: Language Models do not Recognize Identifier Swaps in Python

Link- https://arxiv.org/abs/2305.15507

Abstract- Large Language Models (LLMs) have successfully been applied to code generation tasks, raising the question of how well these models understand programming. Typical programming languages have invariances and equivariances in their semantics that human programmers intuitively understand and exploit, such as the (near) invariance to the renaming of identifiers. We show that LLMs not only fail to properly generate correct Python code when default function names are swapped, but some of them even become more confident in their incorrect predictions as the model size increases, an instance of the recently discovered phenomenon of Inverse Scaling, which runs contrary to the commonly observed trend of increasing prediction quality with increasing model size. Our findings indicate that, despite their astonishing typical-case performance, LLMs still lack a deep, abstract understanding of the content they manipulate, making them unsuitable for tasks that statistically deviate from their training data, and that mere scaling is not enough to achieve such capability.

My post on it

Authors- Antonio Valerio Miceli Barone, Fazl Barez, Ioannis Konstas, Shay Cohen.

2) Tree of Thoughts: Deliberate Problem Solving with Large Language Models

Paper: https://arxiv.org/abs/2305.10601

Abstract: Language models are increasingly being deployed for general problem solving across a wide range of tasks, but are still confined to token-level, left-to-right decision-making processes during inference. This means they can fall short in tasks that require exploration, strategic lookahead, or where initial decisions play a pivotal role. To surmount these challenges, we introduce a new framework for language model inference, Tree of Thoughts (ToT), which generalizes over the popular Chain of Thought approach to prompting language models, and enables exploration over coherent units of text (thoughts) that serve as intermediate steps toward problem solving. ToT allows LMs to perform deliberate decision making by considering multiple different reasoning paths and self-evaluating choices to decide the next course of action, as well as looking ahead or backtracking when necessary to make global choices. Our experiments show that ToT significantly enhances language models’ problem-solving abilities on three novel tasks requiring non-trivial planning or search: Game of 24, Creative Writing, and Mini Crosswords. For instance, in Game of 24, while GPT-4 with chain-of-thought prompting only solved 4% of tasks, our method achieved a success rate of 74%.

Authors- Shunyu Yao, Dian Yu, Jeffrey Zhao, Izhak Shafran, Thomas L. Griffiths, Yuan Cao, Karthik Narasimhan

My recommendation when tackling this paper would be to watch Yannic Kilcher’s excellent breakdown over here-

3) Is Avoiding Extinction from AI Really an Urgent Priority?

Link- https://aisnakeoil.substack.com/p/is-avoiding-extinction-from-ai-really

Authors- Seth Lazar, Jeremy Howard, and Arvind Narayanan.

Great writeup on the dangers of focusing on existential risks of AI and forgetting the other dangers. Love this particular passage-

“Indeed, focusing on this particular threat might exacerbate the more likely risks. The history of technology to date suggests that the greatest risks come not from technology itself, but from the people who control the technology using it to accumulate power and wealth. The AI industry leaders who have signed this statement are precisely the people best positioned to do just that. And in calling for regulations to address the risks of future rogue AI systems, they have proposed interventions that would further cement their power. We should be wary of Prometheans who want to both profit from bringing the people fire, and be trusted as the firefighters.”

4) The AI Healthcare Report: 5/26/23

Link- https://www.linkedin.com/pulse/ai-healthcare-report-52623-dylan-reid-moskowitz-/

Good resource if you’d like to track some of the advancements in the space of legislation for AI in healthcare.

Author- Dylan Reid(Moskowitz)

5) A rant about why technologists are bad at predictions

Link- https://twitter.com/KevinAFischer/status/1662853371118641154

Hinton infamously predicted that AI would replace radiologists (not dissimilar to how Musk promised Full Self-Driving by 2020). And neither prediction worked out. So why do prominent technologists get predictions so wrong?

Kevin Fischer has a great thread on why the Radiologists never got replaced, and why so many predictions on AI have been off. One of my favorite quotes is- “thinkers have a pattern where they are so divorced from implementation details that applications seem trivial, when in reality, the small details are exactly where value accrues.”

6) What they Don’t tell you about A.I. Jobs Disruption?

Link- https://aisupremacy.substack.com/p/what-they-dont-tell-you-about-ai

Much has been made about the hype of AI replacing people’s jobs. While most of it is hype, there are some jobs — think typists, telemarketers, etc, which have a higher likelihood of being replaced. Unfortunately, much of the work being done here is female-dominated, making them vulnerable to replacement. This has the potential to deepen gender inequality

Author- Michael Spencer

7) LIMA: Less Is More for Alignment

Link- https://arxiv.org/abs/2305.11206

Abstract- Large language models are trained in two stages: (1) unsupervised pretraining from raw text, to learn general-purpose representations, and (2) large scale instruction tuning and reinforcement learning, to better align to end tasks and user preferences. We measure the relative importance of these two stages by training LIMA, a 65B parameter LLaMa language model fine-tuned with the standard supervised loss on only 1,000 carefully curated prompts and responses, without any reinforcement learning or human preference modeling. LIMA demonstrates remarkably strong performance, learning to follow specific response formats from only a handful of examples in the training data, including complex queries that range from planning trip itineraries to speculating about alternate history. Moreover, the model tends to generalize well to unseen tasks that did not appear in the training data. In a controlled human study, responses from LIMA are either equivalent or strictly preferred to GPT-4 in 43% of cases; this statistic is as high as 58% when compared to Bard and 65% versus DaVinci003, which was trained with human feedback. Taken together, these results strongly suggest that almost all knowledge in large language models is learned during pretraining, and only limited instruction tuning data is necessary to teach models to produce high quality output.

Found this paper on Davis Blalock’s amazing newsletter over here (or click below).

Authors- Chunting Zhou, Pengfei Liu, Puxin Xu, Srini Iyer, Jiao Sun, Yuning Mao, Xuezhe Ma, Avia Efrat, Ping Yu, Lili Yu, Susan Zhang, Gargi Ghosh, Mike Lewis, Luke Zettlemoyer, Omer Levy

Other interesting Reads

1) Unlocking the Potential of Hybrid Work

Abi’s newsletter delivers hit after hit for software engineering and tech. His most recent writeup on hybrid work is another certified classic if you’re into leadership

Link- https://newsletter.abinoda.com/p/hybrid-work-productivity

Author- Abi Noda

2) Native language shapes brain wiring

Scientists at the Max Planck Institute for Human Cognitive and Brain Sciences in Leipzig have found evidence that the language we speak shapes the connectivity in our brains that may underlie the way we think. With the help of magnetic resonance tomography, they looked deep into the brains of native German and Arabic speakers and discovered differences in the wiring of the language regions in the brain.

Link- https://www.mpg.de/20008844/our-native-language-shapes-the-brain-wiring

Authors- Max Plank Institute Berlin

Photo by Jon Tyson on Unsplash

Cool Vids-

Low-rank Adaption of Large Language Models: Explaining the Key Concepts Behind LoRA- Chris Alexiuk

Don’t Underestimate The German Economy | Economics Explained-

Why Julian Nagelsmann is the perfect Real Madrid manager- Tifo Football

I’ll catch y’all with more of these next week. In the meanwhile if you’d like to find me, here are my social links-

Reach out to me

Use the links below to check out my other content, learn more about tutoring, reach out to me about projects, or just to say hi.

Small Snippets about Tech, AI and Machine Learning over here

AI Newsletter- https://artificialintelligencemadesimple.substack.com/

My grandma’s favorite Tech Newsletter- https://codinginterviewsmadesimple.substack.com/

Check out my other articles on Medium. : https://rb.gy/zn1aiu

My YouTube: https://rb.gy/88iwdd

Reach out to me on LinkedIn. Let’s connect: https://rb.gy/m5ok2y

My Instagram: https://rb.gy/gmvuy9

My Twitter: https://twitter.com/Machine01776819

BECOME a WRITER at MLearning.ai // AI Factory XR Super Cheap AI

--

--

Devansh

Writing about AI, Math, the Tech Industry and whatever else interests me. Join my cult to gain inner peace and to support my crippling chocolate milk addiction