205 private links
This course focuses on efficient machine learning and systems. This is a crucial area as deep neural networks demand extraordinary levels of computation, hindering its deployment on everyday devices and burdening the cloud infrastructure. This course introduces efficient AI computing techniques that enable powerful deep learning applications on resource-constrained devices. Topics include model compression, pruning, quantization, neural architecture search, distributed training, data/model parallelism, gradient compression, and on-device fine-tuning. It also introduces application-specific acceleration techniques for large language models and diffusion models. Students will get hands-on experience implementing model compression techniques and deploying large language models (Llama2-7B) on a laptop.
using a thunderbolt 5 bridge and https://github.com/mit-han-lab/TinyChatEngine
ai speaks with itself and reveals the mad dreams of an electric mind.
interview with David Lee, the CCO of Squarespace, poses an interesting question: Will AI make human creativity a luxury?
Learn how to build an agent that can reason over your documents and answer complex questions. Learn from the co-founder and CEO of LlamaIndex
Roie Schwaber-Cohen, Staff Developer Advocate at Pinecone, joins Ben and Ryan to break down what retrieval-augmented generation (RAG) is and why the concept is central to the AI conversation.
Volvo AD
Andrej Karpathy - The Tokenizer is a necessary and pervasive component of Large Language Models (LLMs), where it translates between strings and tokens (text chunks).
Lucian Grainge, the chairman of UMG, has helped record labels rake in billions of dollars from streaming. Can he do the same with generative artificial intelligence? John Seabrook reports.
A complete GPT2 implementation as a single SQL query in PostgreSQL.
AI coding assistants are here to stay—but just how big a difference they make is still unclear.
Mozilla’s innovation group and Justine Tunney just released llamafile, and I think it’s now the single best way to get started running Large Language Models,
Draw a picture (fast) with tldraw with https://www.fal.ai
The New Yorker - James Somers, a professional coder, writes about the astonishing scripting skills of A.I. chatbots like GPT-4 and considers the future of a once exalted craft. ~ "What I learned was that programming is not really about knowledge or skill but simply about patience, or maybe obsession. Programmers are people who can endure an endless parade of tedious obstacles."
A world of AI-assisted writing and reviewing might transform the nature of the scientific paper.
Harvard Business Publishing Education
browser-based search engine for Wikipedia, where you can search more abstractly
Digging into the philosophical roots of the battle between Essentialists and Pragmatists
on Github CoPilot
The model beyond ChatGPT passed the first three sommelier theory exams. What does that mean for the wine business and our professional certifications?
LLM that'll fit on a small device with low memory
3/23
An influential M.I.T. professor and an outside-the-box scientific theorist, he gained fame with unorthodox views as a pioneer in digital physics.
Large language models (LLMs) have only just emerged into mainstream thought, and already they’ve shown themselves to be a powerful tool for interacting with data. While some might classify them as merely a really cool new form of UI, others think that this may be the start of artificial general intelligence.
Much has been written about the Reddit boycott recently, but I have kind of a wild take. what if we thought of Reddit as, functionally, subservient to OpenAI?
What would a copilot for writing and thinking look like? To try answering this question, I built a prototype: Obsidian-Copilot
In a now-taken-down blog post summarizing an event with Sam Altman, Altman revealed that he doesn’t believe that ChatGPT plugins have product-market fit (outside of the browsing plugin) and won’t be coming to the API soon. Why? A few hypotheses (not mutually exclusive). "Chat is not the right UX for plugins. If you know what you want to do, it’s often easier to just do a few clicks on the website. If you don’t, just a chat interface makes it hard to steer the model toward your goal."
Clinical predictive models can help physicians and administrators make decisions by forecasting clinical and operational events.
In March and April 2023, the Python Software Foundation (PSF) received three (3) subpoenas for PyPI user data from the The Python Package Index.
SwingVision is the #1 tennis app for real-time automated score keeping, stats & line calling. Measure your serve speed, match stats & more!
Amelia Wattenberger makes a convincing argument for why chatbots are a terrible interface for LLMs.
Introducing MPT-7B, the latest entry in our MosaicML Foundation Series. MPT-7B is a transformer trained from scratch on 1T tokens of text and code. It is open source, available for commercial use, and matches the quality of LLaMA-7B. MPT-7B was trained on the MosaicML platform in 9.5 days with zero human intervention at a cost of ~$200k. Starting today, you can train, finetune, and deploy your own private MPT models, either starting from one of our checkpoints or training from scratch. For inspiration, we are also releasing three finetuned models in addition to the base MPT-7B: MPT-7B-Instruct, MPT-7B-Chat, and MPT-7B-StoryWriter-65k+, the last of which uses a context length of 65k tokens!
With the MosaicBERT architecture + training recipe, you can now pretrain a competitive BERT-Base model from scratch on the MosaicML platform for $20. We’ve released the pretraining and finetuning code, as well as the pretrained weights.
Good basic explainer on vector databases.
Sam Altman is the CEO of OpenAI, the company behind GPT-4, ChatGPT, DALL-E, Codex, and many other state-of-the-art AI technologies.
Use the new GPT-4 api to build a chatGPT chatbot for multiple Large PDF files.
This time, we spent the whole episode talking about large language models: ChatGPT, GPT-4, Bing, Bard, Claude, LLaMA and more.
Long before the invention of the general-purpose computer, bureaucrats and researchers had begun gathering and cross-tabulating sets of numbers about populations—heights, weights, ages, sexes, races, political parties, incomes—using punch cards and tabulating machines. Sergey Brin said "Auto insurance companies analyse accident data and set insurance rates of individuals according to age, gender, vehicle type,” he pointed out. “If they were allowed to by law, they would also use race, religion, handicap, and any other attributes they find are related to accident rate.”
Talk with an LLaMA AI in your terminal / Port of OpenAI's Whisper model in C/C++. Contribute to ggerganov/whisper.cpp development by creating an account on GitHub.
The New Yorker | OpenAI’s chatbot offers paraphrases, whereas Google offers quotes. Which do we prefer?
Twitter thread from Jackson Fall on using gpt to create a website for making money.
We report the development of GPT-4, a large-scale, multimodal model which can accept image and text inputs and produce text outputs. While less capable than humans in many real-world scenarios, GPT-4 exhibits human-level performance on various professional and academic benchmarks, including passing a simulated bar exam with a score around the top 10% of test takers.
Dataset Repository of Awesome ChatGPT Prompts from https://github.com/f/awesome-chatgpt-prompts
A creature is formed of clay. A puppet becomes a boy. A monster rises in a lab. A computer takes over a spaceship. And all manner of robots serve or control us. For generations we’ve told ourselves stories, using themes of magic and science, about inanimate things that we bring to life or imbue with power beyond human capacity.
Proving you're a human on a web flooded with generative AI content
~ Computer Scientist - Stanford Artificial Intelligence Laboratory... wrote "AI is a fantasy"
The ethics of AI are constantly debated. But does anyone ask the AI?
Project Debater is designed by IBM research. It will deliver a speech based on over 1,100 arguments collected from Union members and others over the past week. It will not be taking points of information.