Making Sense of the Promise (and Risks) of Large Language Models | by TDS Editors | Apr, 2023


  • Machine learning research in a post-ChatGPT world. The inner workings of some of the most ubiquitous LLMs (like OpenAI’s GPT models) remain tightly guarded by corporate entities. What does that mean for NLP researchers whose future projects are likely to hit a proprietary wall sooner or later? Anna Rogers proposes a powerful idea: “That which is not open and reasonably reproducible cannot be considered a requisite baseline.”
  • Assessing the impact of recent advances in conversational AI. The size and power of LLMs have made it possible for human-machine interaction to take a giant leap in mere months. Gadi Singer reflects on how we got here—and on what is still missing before the machines start communicating in a human-like manner.
  • Welcome to a new ethical (and practical) quagmire. LLMs require massive amounts of training data, which has turned out to be a major vector of risk. These models are currently facing intense scrutiny for the potential presence of biased, private, or harmful text and the unauthorized inclusion of copyrighted material. What are users to do in the meantime? Lingjuan Lyu’s debut TDS article presents a thoughtful overview of the stakes and dangers we should all be aware of if we’re to produce AI-generated content responsibly.
Photo by Libby Penner on Unsplash
  • The wide horizon of LLM-powered tools. Just a few months ago, the LangChain library was flying under most practitioners’ radars. Well, not anymore: it’s become a go-to resource for many tinkerers who want to leverage LLMs to build new apps. Dr. Varshita Sher’s latest deep dive is a helpful, hands-on introduction to the library’s core building blocks.
  • Identifying drift and detecting anomalies with LLMs. As the novelty of generating clunky poems with ChatGPT fades, fresh use cases continue to emerge — and many of them might end up streamlining data science workflows. Case in point: Aparna Dhinakaran, Jason Lopatecki, and Christopher Broan’s latest post, which outlines a promising approach for using using LLM embeddings for anomaly and drift detection.
  • Bonus read: go one level deeper. If LLMs make user-facing applications like ChatGPT possible, transformer neural networks are the architecture that made LLMs possible in the first place. To get your bearings around this crucial (and often complex) topic, explore Soran Ghaderi’s detailed “map” of past and current transformers research.


  • Machine learning research in a post-ChatGPT world. The inner workings of some of the most ubiquitous LLMs (like OpenAI’s GPT models) remain tightly guarded by corporate entities. What does that mean for NLP researchers whose future projects are likely to hit a proprietary wall sooner or later? Anna Rogers proposes a powerful idea: “That which is not open and reasonably reproducible cannot be considered a requisite baseline.”
  • Assessing the impact of recent advances in conversational AI. The size and power of LLMs have made it possible for human-machine interaction to take a giant leap in mere months. Gadi Singer reflects on how we got here—and on what is still missing before the machines start communicating in a human-like manner.
  • Welcome to a new ethical (and practical) quagmire. LLMs require massive amounts of training data, which has turned out to be a major vector of risk. These models are currently facing intense scrutiny for the potential presence of biased, private, or harmful text and the unauthorized inclusion of copyrighted material. What are users to do in the meantime? Lingjuan Lyu’s debut TDS article presents a thoughtful overview of the stakes and dangers we should all be aware of if we’re to produce AI-generated content responsibly.
Photo by Libby Penner on Unsplash
  • The wide horizon of LLM-powered tools. Just a few months ago, the LangChain library was flying under most practitioners’ radars. Well, not anymore: it’s become a go-to resource for many tinkerers who want to leverage LLMs to build new apps. Dr. Varshita Sher’s latest deep dive is a helpful, hands-on introduction to the library’s core building blocks.
  • Identifying drift and detecting anomalies with LLMs. As the novelty of generating clunky poems with ChatGPT fades, fresh use cases continue to emerge — and many of them might end up streamlining data science workflows. Case in point: Aparna Dhinakaran, Jason Lopatecki, and Christopher Broan’s latest post, which outlines a promising approach for using using LLM embeddings for anomaly and drift detection.
  • Bonus read: go one level deeper. If LLMs make user-facing applications like ChatGPT possible, transformer neural networks are the architecture that made LLMs possible in the first place. To get your bearings around this crucial (and often complex) topic, explore Soran Ghaderi’s detailed “map” of past and current transformers research.

FOLLOW US ON GOOGLE NEWS

Read original article here

Denial of responsibility! Techno Blender is an automatic aggregator of the all world’s media. In each content, the hyperlink to the primary source is specified. All trademarks belong to their rightful owners, all materials to their authors. If you are the owner of the content and do not want us to publish your materials, please contact us by email – admin@technoblender.com. The content will be deleted within 24 hours.
Ai NewsAprEditorslanguageLargemachine learningMakingModelspromiseRisksSenseTDSTechnoblender
Comments (0)
Add Comment