mathstodon.xyz is one of the many independent Mastodon servers you can use to participate in the fediverse.
A Mastodon instance for maths people. We have LaTeX rendering in the web interface!

Server stats:

2.7K
active users

#languagemodeling

0 posts0 participants0 posts today
Harald Klinke<p>Andrey Markov &amp; Claude Shannon Counted Letters to Build the First Language-Generation Models <br>Shannon’s said: “OCRO HLI RGWR NMIELWIS”<br><a href="https://det.social/tags/Shannon" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Shannon</span></a> <a href="https://det.social/tags/Markov" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Markov</span></a> <a href="https://det.social/tags/NLP" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>NLP</span></a> <a href="https://det.social/tags/AIhistory" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>AIhistory</span></a> <a href="https://det.social/tags/LanguageModeling" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LanguageModeling</span></a><br><a href="https://spectrum.ieee.org/andrey-markov-and-claude-shannon-built-the-first-language-generation-models" rel="nofollow noopener noreferrer" translate="no" target="_blank"><span class="invisible">https://</span><span class="ellipsis">spectrum.ieee.org/andrey-marko</span><span class="invisible">v-and-claude-shannon-built-the-first-language-generation-models</span></a></p>
khushnuma<p>Mastering these core NLP techniques is crucial for any data scientist dealing with text data. From tokenization to language modeling, each method serves a unique purpose in processing, analyzing, and extracting valuable insights from textual information.</p><p><a href="https://mastodon.social/tags/NLP" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>NLP</span></a> <a href="https://mastodon.social/tags/DataScience" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>DataScience</span></a> <a href="https://mastodon.social/tags/Tokenization" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>Tokenization</span></a> <a href="https://mastodon.social/tags/LanguageModeling" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LanguageModeling</span></a> <a href="https://mastodon.social/tags/TextAnalysis" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>TextAnalysis</span></a> <a href="https://mastodon.social/tags/TextMining" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>TextMining</span></a> <a href="https://mastodon.social/tags/MachineLearning" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>MachineLearning</span></a> </p><p>read more: <a href="https://blogulr.com/khushnuma7861/topnlptechniqueseverydatascientistshouldknow-120682" rel="nofollow noopener noreferrer" translate="no" target="_blank"><span class="invisible">https://</span><span class="ellipsis">blogulr.com/khushnuma7861/topn</span><span class="invisible">lptechniqueseverydatascientistshouldknow-120682</span></a></p>
Naomi Saphra<p>New <a href="https://sigmoid.social/tags/languagemodeling" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>languagemodeling</span></a> <a href="https://sigmoid.social/tags/nlp" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>nlp</span></a> <a href="https://sigmoid.social/tags/ai" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>ai</span></a> <a href="https://sigmoid.social/tags/paper" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>paper</span></a>, led by Angelica Chen! We break the steepest MLM training loss drop into *2* phase changes: first in internal grammatical structure, then external capabilities. Big implications for emergence, simplicity bias, and interpretability! <a href="https://arxiv.org/abs/2309.07311" rel="nofollow noopener noreferrer" target="_blank"><span class="invisible">https://</span><span class="">arxiv.org/abs/2309.07311</span><span class="invisible"></span></a></p>
Victoria Stuart 🇨🇦 🏳️‍⚧️<p>Revealing the structure of language model capabilities<br><a href="https://arxiv.org/abs/2306.10062" rel="nofollow noopener noreferrer" target="_blank"><span class="invisible">https://</span><span class="">arxiv.org/abs/2306.10062</span><span class="invisible"></span></a></p><p>Building a theoretical understanding of the capabilities of large language models (LLMs) is vital for our ability to predict &amp; explain the behavior of these systems. ... we analyzed data from 29 LLMs / 27 cognitive tasks. LLM are better explained by 3 well-delineated factors that represent reasoning, comprehension &amp; core language modeling.</p><p><a href="https://mastodon.social/tags/LLM" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LLM</span></a> <a href="https://mastodon.social/tags/LargeLanguageModels" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LargeLanguageModels</span></a> <a href="https://mastodon.social/tags/reasoning" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>reasoning</span></a> <a href="https://mastodon.social/tags/LanguageModeling" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LanguageModeling</span></a> <a href="https://mastodon.social/tags/comprehension" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>comprehension</span></a> <a href="https://mastodon.social/tags/GPT" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>GPT</span></a></p>
Netherlands eScience Center<p><a href="https://akademienl.social/tags/LanguageModeling" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>LanguageModeling</span></a> is trending, to a large extent because of <a href="https://akademienl.social/tags/ChatGPT" class="mention hashtag" rel="nofollow noopener noreferrer" target="_blank">#<span>ChatGPT</span></a>. But did you know language modeling has been with us for more than a century? And that it was born of the collaboration of a poet and a mathematician? </p><p>Our engineer Carsten Schnober tells us more:<br><a href="https://blog.esciencecenter.nl/language-modeling-the-first-100-years-357556816148" rel="nofollow noopener noreferrer" target="_blank"><span class="invisible">https://</span><span class="ellipsis">blog.esciencecenter.nl/languag</span><span class="invisible">e-modeling-the-first-100-years-357556816148</span></a></p>