eduroam-prg-gm-1-3-245.net.univ-paris-diderot.fr 2025-9-19:10:2:53
This commit is contained in:
		
							
								
								
									
										2
									
								
								.obsidian/graph.json
									
									
									
									
										vendored
									
									
								
							
							
						
						
									
										2
									
								
								.obsidian/graph.json
									
									
									
									
										vendored
									
									
								
							@@ -130,6 +130,6 @@
 | 
			
		||||
  "repelStrength": 5.263671875,
 | 
			
		||||
  "linkStrength": 1,
 | 
			
		||||
  "linkDistance": 30,
 | 
			
		||||
  "scale": 2.422433756553115,
 | 
			
		||||
  "scale": 0.15541659193590038,
 | 
			
		||||
  "close": true
 | 
			
		||||
}
 | 
			
		||||
@@ -15,6 +15,10 @@ $x_{i} \in \mathbb{R}$ is a scalar
 | 
			
		||||
 | 
			
		||||
one-hot : boolean vector with all zeroes but one value. Usefull if each dimension represents a word of the vocabulary
 | 
			
		||||
 | 
			
		||||
BOW : Bag Of Words
 | 
			
		||||
You could represent sentences like that :
 | 
			
		||||
Let our vocabulary be : `V = 'le' 'un' 'garcon' 'lit' 'livre' 'regarde'`
 | 
			
		||||
Then "le garcon lit le livre" would be written by counting the number of occurences of each word of the sentence in a vector, so `2 0 1 1 1 0` (the formula is )
 | 
			
		||||
Then "le garcon lit le livre" would be written by counting the number of occurences of each word of the sentence in a vector, so `2 0 1 1 1 0` (the formula is `sentence +⌿⍤(∘.≡) vocabulary`)
 | 
			
		||||
 | 
			
		||||
$\cos(u, v) = \frac{u\cdot v}{\|u\| \| v\|}$
 | 
			
		||||
 | 
			
		||||
 
 | 
			
		||||
		Reference in New Issue
	
	Block a user