Topics Top articles New articles Updated articles Top users New users New discussions Top discussions New comments+ New article
Interesting layer skip architecture thing.
Apparently destroyed ImageNet 2015 and became very very famous as such.
CNN convolution kernels are also learnt by Ciro Santilli 35 Updated 2025-01-10 +Created 1970-01-01
CNN convolution kernels are not hardcoded. They are learnt and optimized via backpropagation. You just specify their size! Example in PyTorch you'd do just:as used for example at: activatedgeek/LeNet-5.
nn.Conv2d(1, 6, kernel_size=(5, 5))
This can also be inferred from: stackoverflow.com/questions/55594969/how-to-visualise-filters-in-a-cnn-with-pytorch where we see that the kernels are not perfectly regular as you'd expected from something hand coded.
Deep learning is the name artificial neural networks basically converged to in the 2010s/2020s.
It is a bit of an unfortunate as it suggests something like "deep understanding" and even reminds one of AGI, which it almost certainly will not attain on its own. But at least it sounds good.
Ciro Santilli's favorites, including album when they're more of a one hit wonder:
- Allan Holdsworth
- Joe Satriani
- Jean-Luc Ponty
- Chick Corea
- Billy Cobham
- Pat Metheny
- Keith Jarrett (Arbor Zena)
- John Abercrombie (Timelesss)
- Bill Brisell (Blues Dream)
- Larry Coryell (Spaces)
Many good albums, Ciro Santilli's favorites:
However, there is nothing in the immediate definition that prevents us from having a ring instead, i.e. a field but without the commutative property and inverse elements.
The only thing is that then we would need to differentiate between different orderings of the terms of multivariate polynomial, e.g. the following would all be potentially different terms:while for a field they would all go into a single term:so when considering a polynomial over a ring we end up with a lot more more possible terms.
If the ring is a commutative ring however, polynomials do look like proper polynomials: Section "Polynomial over a commutative ring".
An imagenet10 subset by fast.ai.
Size of full sized image version: 1.5 GB.
This is the one used on MLperf v2.1 ResNet, likely one of the most popular choices out there.
2017 challenge subset:
- train: 118k images, 18GB
- validation: 5k images, 1GB
- test: 41k images, 6GB
Interesting website, hosts mostly:
- datasets
- ANN models
- some live running demos called "apps": e.g. huggingface.co/spaces/ronvolutional/ai-pokemon-card
What's the point of this website vs GitHub? www.reddit.com/r/MLQuestions/comments/ylf4be/whats_the_deal_with_hugging_faces_popularity/
They have a tutorial at: jena.apache.org/tutorials/sparql.html
Once you've done the Apache Jena CLI tools setup we can query all users with Full Name (FN) "John Smith" directly fom the rdf/vcard.ttl Turtle RDF file with the rdf/vcard.rq SPARQL query:and that outputs:
sparql --data=rdf/vcard.ttl --query=rdf/vcard.rq
---------------------------------
| x |
=================================
| <http://somewhere/JohnSmith/> |
---------------------------------
Within the The Holy Trinity of popular Brazilian music, Caetano has the most New Age religious feel to him. He is also perhaps the most varied of the trinity however, also covering heavier topics at times.
Groups concepts by hyponymy and hypernymy and meronymy and holonymy. That actually makes a lot of sense! TODO: is there a clear separation between hyponymy and meronymy?
Does not contain intermediat scientific terms, only very common ones, e.g. no mention, of "Josephson effect", "photoelectric effect"
Pinned article: ourbigbook/introduction-to-the-ourbigbook-project
Welcome to the OurBigBook Project! Our goal is to create the perfect publishing platform for STEM subjects, and get university-level students to write the best free STEM tutorials ever.
Everyone is welcome to create an account and play with the site: ourbigbook.com/go/register. We belive that students themselves can write amazing tutorials, but teachers are welcome too. You can write about anything you want, it doesn't have to be STEM or even educational. Silly test content is very welcome and you won't be penalized in any way. Just keep it legal!
We have two killer features:
- topics: topics group articles by different users with the same title, e.g. here is the topic for the "Fundamental Theorem of Calculus" ourbigbook.com/go/topic/fundamental-theorem-of-calculusArticles of different users are sorted by upvote within each article page. This feature is a bit like:
- a Wikipedia where each user can have their own version of each article
- a Q&A website like Stack Overflow, where multiple people can give their views on a given topic, and the best ones are sorted by upvote. Except you don't need to wait for someone to ask first, and any topic goes, no matter how narrow or broad
This feature makes it possible for readers to find better explanations of any topic created by other writers. And it allows writers to create an explanation in a place that readers might actually find it. - local editing: you can store all your personal knowledge base content locally in a plaintext markup format that can be edited locally and published either:This way you can be sure that even if OurBigBook.com were to go down one day (which we have no plans to do as it is quite cheap to host!), your content will still be perfectly readable as a static site.
- to OurBigBook.com to get awesome multi-user features like topics and likes
- as HTML files to a static website, which you can host yourself for free on many external providers like GitHub Pages, and remain in full control
- Internal cross file references done right:
- Infinitely deep tables of contents:
All our software is open source and hosted at: github.com/ourbigbook/ourbigbook
Further documentation can be found at: docs.ourbigbook.com
Feel free to reach our to us for any help or suggestions: docs.ourbigbook.com/#contact