Discussion
Loading...

#Tag

Log in
  • About
  • Code of conduct
  • Privacy
  • Users
  • Instances
  • About Bonfire
David Adler
David Adler
@perspektivbrocken@social.tchncs.de  ·  activity timestamp 18 hours ago

Kennt jemand gutes Bildungsmaterial um mit einer Gruppe selbst ein kleines, sehr simples #LLM zu bauen? Ziel wäre, dir Basics etwas zu entzaubern. Zielgruppe in dem Fall Erwachsene, aber ohne irgendwelche Vorkenntnisse in dem Bereich. Vielleicht @fedilz_infos ?

Andreas Wagner
Andreas Wagner
@anwagnerdreas@hcommons.social replied  ·  activity timestamp 12 hours ago

@perspektivbrocken

@thomasrenkert weiß bestimmt etwas zu berichten, der arbeitet in der Lehrerbildung (richtig?) und hat den ParzivAI Assistenten entwickelt: https://agki-dh.github.io/pages/webinar/page-9.html / https://hse.hypotheses.org/6066.

An Raschkas Building LLMs from Scratch hab ich auch schon gedacht, das wurde ja schon erwähnt. Es gibt im Netz Code dazu (https://github.com/rasbt/LLMs-from-scratch) und ein paar Leute, die ihre Erfahrungen damit/darum herum posten, z.B. https://www.gilesthomas.com/llm-from-scratch . https://huggingface.co/blog/gszauer/minimal-llm und https://readmedium.com/how-to-build-an-llm-from-scratch-8c477768f1f9 sind auch gut. Alles Englisch halt.

Wenn es nicht so sehr ums selber bauen sondern auch um Erklärungen geht, sind glaube ich die Videos von Andrej Karpathy (z.B. https://youtu.be/7xTGNNLPyMI?si=-YlKsMGuBnW5GO44 oder https://youtu.be/zduSFxRajkE?si=M-cYkZVX6N8M-Oxy ) oder von Thomas Wolf (https://youtu.be/2-SPH9hIKT8?si=IV-hyIaYCYi-dQTZ) ganz gut. https://youtu.be/LPZh9BOjkQs?si=uxW_igUPN92C6wg2 auch.

Und ich finde ja so "Simulationen" auch sehr hilfreich: https://www.soekia.ch/gpt.html / https://bbycroft.net/llm / https://poloclub.github.io/transformer-explainer/

#LLM #AI #Explainer #Transformer #GPT #BERT

Transformer Explainer: LLM Transformer Model Visually Explained

An interactive visualization tool showing you how transformer models work in large language models (LLM) like GPT.

LLM Visualization

A 3D animated visualization of an LLM with a walkthrough.

SoekiaGPT - Das didaktische Sprachmodell

SoekiaGPT ist ein Textgenerator speziell für den Unterricht. Mit SoekiaGPT kannst Du hinter die Kulissen schauen und damit einige Grundprinzipien von Textgeneratoren wie ChatGPT kennenlernen.

How to Build an LLM from Scratch

Data Curation, Transformers, Training at Scale, and Model Evaluation

Build a simple LLM from scratch

A Blog post by Gabor Szauer on Hugging Face

LLM from scratch :: Giles' blog

(
GitHub

GitHub - rasbt/LLMs-from-scratch: Implement a ChatGPT-like LLM in PyTorch from scratch, step by step

Implement a ChatGPT-like LLM in PyTorch from scratch, step by step - rasbt/LLMs-from-scratch
Fokus Lehrerbildung

KI Showcase: Der Chatbot “ParzivAI”

Am 22.07.24 stellten Dr. Florian Nieser und Dr. Thomas Renkert die Alpha-Version des von ihnen entwickelten Chatbots ParzivAI vor. ParzivAI ist auf die Vermittlung der mittelhochdeutschen Sprache und Geschichtsdidaktik des Mittelalters spezialisiert und besteht aus einem Sprachmodell (LLM), einem RAG-System für Dokumentenzugriff, sowie einem Frontend mit Feedback-Funktion

DHd-AG Angewandte Generative KI in den Digitalen Geisteswissenschaften (AGKI-DH)

The Applied Generative AI in Digital Humanities (AGKI-DH) working group focuses on exploring and showcasing the applications of generative AI, especially Large Language Models (LLMs), in digital humanities. This includes both commercial and open-source models, recognizing the significant advancements and the widespread discussion on AI issues like bias, hallucination, high computational footprint, data security, and potential destructive capabilities. Amidst these challenges, the group aims to harness the opportunities in AI development for digital humanities, fostering an ongoing exchange on core AI concepts and practical experiences. It seeks collaboration with other groups on greening (AI footprint), digital humanities theory, and empowerment (ethical aspects). The goals include identifying, assessing, and categorizing generative AI applications in digital humanities, establishing best practices, developing evaluation criteria, facilitating discussions and knowledge exchange, and balancing opportunities and risks.
  • Copy link
  • Flag this comment
  • Block
Hacker News
Hacker News
@h4ckernews@mastodon.social  ·  activity timestamp 3 weeks ago

Starting from scratch: Training a 30M Topological Transformer

https://www.tuned.org.uk/posts/013_the_topological_transformer_training_tauformer

#HackerNews #Starting #from #scratch: #Training #a #30M #Topological #Transformer

topologicaltransformer #training #machinelearning #AI #research #deeplearning

013_the_topological_transformer_training_tauformer - AI Research Engineering

Dash Remover
Dash Remover
@dashremover@mastodon.social replied  ·  activity timestamp 3 weeks ago

Incredible how we went from 'Training a 30M Topological Transformer' to 'Training #a #30M #Topological #Transformer' like some kind of SEO summoning ritual. Next stop: bake the weights into sourdough and post on Medium. #AI 🍞🤖

  • Copy link
  • Flag this comment
  • Block
Hacker News
Hacker News
@h4ckernews@mastodon.social  ·  activity timestamp 3 weeks ago

Starting from scratch: Training a 30M Topological Transformer

https://www.tuned.org.uk/posts/013_the_topological_transformer_training_tauformer

#HackerNews #Starting #from #scratch: #Training #a #30M #Topological #Transformer

topologicaltransformer #training #machinelearning #AI #research #deeplearning

013_the_topological_transformer_training_tauformer - AI Research Engineering

  • Copy link
  • Flag this post
  • Block
petersuber
petersuber
@petersuber@fediscience.org  ·  activity timestamp 3 months ago

"Databricks co-founder argues US must go #OpenSource to beat China in AI."
https://techcrunch.com/2025/11/14/databricks-co-founder-argues-us-must-go-open-source-to-beat-china-in-ai/

"Major #AI labs, including OpenAI, Meta, and Anthropic, continue to innovate significantly, yet their innovations remain largely proprietary… #Konwinski argued that for ideas to truly flourish, they need to be freely exchanged and discussed with the larger academic community. He pointed out that generative AI emerged as a direct result of the #Transformer architecture, a pivotal training technique introduced in a freely available research paper [ #OpenAccess in #arXiv]."

  • Copy link
  • Flag this post
  • Block

bonfire.cafe

A space for Bonfire maintainers and contributors to communicate

bonfire.cafe: About · Code of conduct · Privacy · Users · Instances
Bonfire social · 1.0.2-alpha.7 no JS en
Automatic federation enabled
Log in
  • Explore
  • About
  • Members
  • Code of Conduct