You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: config.yaml
+15
Original file line number
Diff line number
Diff line change
@@ -348,6 +348,21 @@ params:
348
348
- icon: fas fa-blog
349
349
url: /blogs/attention-is-all-you-need
350
350
351
+
- title: "PaliGemma: A versatile 3B VLM for transfer"
352
+
content: Paper implementation of Paligemma inspired by Umar Jamil. Currently a work in progress
353
+
image: /projects/paligemma.png
354
+
# featured:
355
+
# name: Demo # Fixit
356
+
# link: https://hugo-profile.netlify.app
357
+
badges:
358
+
- "VLM"
359
+
- "Code Follow Along"
360
+
- "Vision Model"
361
+
- "Transformers"
362
+
links:
363
+
- icon: fab fa-github
364
+
url: https://github.com/sarthak247/PaliGemma
365
+
351
366
- title: 🐱 NekoPDF 📖 - Document Interaction with LLM
352
367
content: |-
353
368
A web app to seamlessly interact with your documents using a Large Language Model over the Langchain framework. Currently support OpenAI and a single PDF file at a time but an option for other LLM backends and multiple PDF files will be added soon.
Copy file name to clipboardExpand all lines: public/blogs/word-embeddings/index.html
+3-6
Original file line number
Diff line number
Diff line change
@@ -547,16 +547,13 @@ <h2 id="tf-idf">TF-IDF</h2>
547
547
<li><strong>Term Frequency (TF)</strong>: This component measures the frequency of a term (word) within a document. It rewards words that appear frequently within a document. It can be calculated as the ratio of the word’s occurrences in a document to the total word count in that document.
548
548
l</li>
549
549
</ul>
550
-
$$TF(term)=\frac{\text{Number of times term appears in a document}}{\text{Total number of items in the document}}$$
551
-
<p>For example, consider our previous example <code>I love coding and coding is fun.</code> Here, TF(coding) is 1/6 as the word <code>and</code> has been ignored.</p>
550
+
$$TF(term)=\frac{\text{Number of times term appears in a document}}{\text{Total number of items in the document}}$$<p>For example, consider our previous example <code>I love coding and coding is fun.</code> Here, TF(coding) is 1/6 as the word <code>and</code> has been ignored.</p>
552
551
<ul>
553
552
<li><strong>Inverse Document Frequeny</strong>: This component measures the rarity or importance of a word across all the documents. It is the log of the inverse of the document frequency where document frequency tells us the number of documents which contain a particular word.</li>
554
553
</ul>
555
-
$$DF(term)=\frac{\text{Documents containing our term}}{\text{Total number of documents}}$$
556
-
<p>Thus, DF tells us about the proportion of documents which contain our word of interest. Thus, we inverse it to make sure that the more common a word is, example stopwords, the less score it gets and a logarithm is taken to dampen or reduce it’s effect on the final calculation.</p>
554
+
$$DF(term)=\frac{\text{Documents containing our term}}{\text{Total number of documents}}$$<p>Thus, DF tells us about the proportion of documents which contain our word of interest. Thus, we inverse it to make sure that the more common a word is, example stopwords, the less score it gets and a logarithm is taken to dampen or reduce it’s effect on the final calculation.</p>
557
555
<!-- raw HTML omitted -->
558
-
$$IDF(term)=\log{\bigg(\frac{\text{Total number of documents}}{\text{Documents containing our term}}\bigg)}$$
559
-
<p>Thus, Inverse Document Frequency (IDF) is a measure of how unique or significant a word is across a collection of documents. It can be computed as the logarithm of the total number of documents divided by the number of documents in which the word occurs, effectively quantifying the word’s rarity and importance in the entire document collection.</p>
556
+
$$IDF(term)=\log{\bigg(\frac{\text{Total number of documents}}{\text{Documents containing our term}}\bigg)}$$<p>Thus, Inverse Document Frequency (IDF) is a measure of how unique or significant a word is across a collection of documents. It can be computed as the logarithm of the total number of documents divided by the number of documents in which the word occurs, effectively quantifying the word’s rarity and importance in the entire document collection.</p>
0 commit comments