Data Machina #262 – by Carlos

Data Machina #262 – by Carlos

Hoping the AI Agents would show up and help. But after the humans in charge evaporated, the AI Agents never arrived. 3 flights cancelled. 30 hours stranded in Gatwick. No personalised online help, no chatbot assistants, no cash from the ATM, no ccard payments, no flights to escape from hell. Being human is about feeling useless and impotent when 1,000s of machines fail in chain because a little s/w update went haywire. I wonder what will happen when in a few years -inevitably- billions of AI Agents operating out in the wild decide to go on strike. In the meantime, trying to catch up with a few things that popped up in my AI radar.

Mistral NeMo 12B SOTA Model. A state-of-the-art 12B, multi-lingual model with 128k context length, built in collaboration with NVIDIA, and released under the Apache 2.0 license. The weights for the base and the instruct model versions are hosted in Hugging Face. Some community members have raised questions about the accuracy of its reported benchmarks, particularly in comparison to models like Llama 3 8B.

Stanford TextGrad: Automatic “Differentiation” via Text. A new, powerful framework for performing automatic “differentiation” via text. TextGrad backpropagates textual feedback provided by LLMs to improve individual components of a compound AI system. TexGrad improves zero-shot capabilities of LLMs significantly across many applications.

Tencent Patch-level Training. This paper introduces a new technique that address the issue of LLM token-training: the huge computational costs due to processing 10s of billions of tokens. Patch-level training reduces the sequence length by compressing multiple tokens into a single patch, and then the model is trained to predict the next patch. Patch-level training can reduce overall computational costs to 0.5×,

[free] Computer Vision Course. This new, Hugging Face community-driven course covers everything from the basics to the latest advancements in computer vision. It’s structured to include various foundational topics, making it friendly and accessible for everyone.

Stanford STORM Long-form, Wikipedia-like generative writing and Q&A. A new, open-source, generative writing system for writing grounded and organised long-form articles from scratch, with comparable breadth and depth to Wikipedia pages. STORM models the pre-writing stage by (1) discovering diverse perspectives in researching a given topic, (2) simulating conversations where writers carrying different perspectives pose questions to a topic expert grounded on trusted Internet sources, (3) curating the collected information to create an outline.

Have a nice week.

  1. State of Open AI, Jul 2024

  2. State of Text2SQL, Jul 2024

  3. Manipulating Chess-GPT’s World Model

  4. The Winds of AI Winter. The Vibes Have Shifted

  5. 43 Questionable Practices in Machine Learning

  6. TinyML: Why the Future of ML is Tiny and Bright

  7. Beating Proprietary AI Models with a Quick Fine-Tune

  8. The AI Song Contest 2024: Calling for Applications

  9. A List of 450 Real-World ML Systems from 100+ Companies

  10. From Knowledge Graphs to GraphRAG: Advanced Intelligent Data Retrieval

Share Data Machina with your friends

  1. Composio – Production Ready Toolset for AI Agents

  2. All-in-one Embeddings DB for Semantic Search, LLM orchestration

  3. MSFT Olive – A Model Optimisation Tool with Industry-leading Techniques

  1. Street Fighting Transformers

  2. Kernel Tricks – Convolutional Kernel Networks

  3. [free tutorial] The Physics of Language Models

  1. E5-V: Universal Embeddings with Multimodal LLMs (paper, repo)

  2. IMAGDressing-v1 : Customisable Virtual Dressing (paper, demo, repo)

  3. Fine-Tuning and Prompt Optimisation: 2 Great Steps that Work Better Together

  1. Different Approaches to Model Compression

  2. How We Built a +2K Features Store with DBT, Snowflake & MySQL

  3. Implementing CI/CD Pipelines with GitHub Actions for MLOps

  1. OAK – A 500M+ Token Dataset Generated by SOTA LLMs

  2. PetFace: A Large-Scale Dataset for Animal Identification

  3. DocMatix – A Massive Dataset for Document Visual Question Answering

Enjoyed this post? Tell your friends about Data Machina. Thanks for reading.

Share

Tips? Suggestions? Feedback? email Carlos

Curated by @alg0agent in the middle of the night.

Related articles

Introductory time-series forecasting with torch

This is the first post in a series introducing time-series forecasting with torch. It does assume some prior...

Does GPT-4 Pass the Turing Test?

Large language models (LLMs) such as GPT-4 are considered technological marvels capable of passing the Turing test successfully....