mathstodon.xyz is one of the many independent Mastodon servers you can use to participate in the fediverse.
A Mastodon instance for maths people. We have LaTeX rendering in the web interface!

Server stats:

2.7K
active users

#imagenet

0 posts0 participants0 posts today

#ConvolutionalNeuralNetworks (#CNNs in short) are immensely useful for many #imageProcessing tasks and much more...

Yet you sometimes encounter some bits of code with little explanation. Have you ever wondered about the origins of the values for image normalization in #imagenet ?

  • Mean: [0.485, 0.456, 0.406] (for R, G and B channels respectively)
  • Std: [0.229, 0.224, 0.225]

Strangest to me is the need for a three-digits precision. Here, after finding the origin of these numbers for MNIST and ImageNet, I am testing if that precision is really important : guess what, it is not (so much) !

👉 if interested in more details, check-out laurentperrinet.github.io/scib

How a stubborn #computerscientist accidentally launched the #deeplearning boom
"You’ve taken this idea way too far," a mentor told Prof. Fei-Fei Li, who was creating a new image #dataset that would be far larger than any that had come before: 14 million images, each labeled with one of nearly 22,000 categories. Then in 2012, a team from Univ of Toronto trained a #neura network on #ImageNet, achieving unprecedented performance in image recognition, dubbed #AlexNet.
arstechnica.com/ai/2024/11/how #AI

Ars Technica · How a stubborn computer scientist accidentally launched the deep learning boomBy Timothy B. Lee

#AI heroic stories and underpaid labour:

"The project was saved when Li learned about Amazon Mechanical Turk, a crowdsourcing platform Amazon had launched a couple of years earlier. "

#Imagenet....How a stubborn computer scientist accidentally launched the deep learning boom - Ars Technica
arstechnica.com/ai/2024/11/how

Ars Technica · How a stubborn computer scientist accidentally launched the deep learning boomBy Timothy B. Lee

🚀 New #AI Research: Simplified Continuous-time Consistency Models (#sCM)

🔬 Key findings:
#OpenAI's new approach matches leading #diffusion models' quality using only 2 sampling steps
• 1.5B parameter model generates samples in 0.11 seconds on single #GPU
• Achieves ~50x wall-clock speedup compared to traditional methods
• Uses less than 10% of typical sampling compute while maintaining quality

🎯 Technical highlights:
• Simplifies theoretical formulation of continuous-time consistency models
• Successfully scaled to 1.5B parameters on #ImageNet at 512×512 resolution
• Demonstrates consistent performance scaling with teacher diffusion models
• Enables real-time generation potential for images, audio, and video

📄 Learn more: openai.com/index/simplifying-s

Replied in thread

@lowd I remember when most ML applications were variations on #MNIST. And #Imagenet, but I only had enough computer at the time to play around with Mnist. But yea, even then "Recommendation Engines" were starting to be the first things anyone mentioned because it was low hanging fruit - something of immediately obvious commercial value with terrific training data and an easy task for deployment.

Can we please, as people who work with large public #datasets, start using torrents? I am just simply trying to find an old version of the #ImageNet Object Detection from Video dataset, and all of the links are broken! For multiple years! Another ImageNet dataset I’m downloading is downloading at 500KB/s.

People have clearly been looking for and using these datasets, and now I need to retrain something and I’m without them. We need to band together and start a torrent tracker for datasets so that we don’t need to rely on one website to download from. With proper permission from the dataset owners of course…

I’m so committed I might buy my own domain and start hosting a torrent tracker. Anyone interested?

#DeepLearning is fun sometimes, especially when you play with #ImageNet...

Here is one result of (our modified version of) ResNet which gives a wrong answer compared to the ground truth label, yet it is visually accurate.

A warning for us all that the objective is not just to reach the highest accuracy, more to better understand what is going on...

👉 This was a result obtained by Emmanuel Daucé from Aix Marseille Université, in a joint work with @jnjer and myself.