this post was submitted on 15 Jul 2024
1968 points (99.6% liked)

Technology

59575 readers
3393 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

Switzerland mandates all software developed for the government be open sourced

Switzerland mandates software source code disclosure for public sector: A legal milestone

https://joinup.ec.europa.eu/collection/open-source-observatory-osor/news/new-open-source-law-switzerland

@[email protected]

#tech #libre

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 2 points 4 months ago (1 children)

That sucks. What was the novel search engine approach?

[–] [email protected] 4 points 4 months ago (1 children)

Using the idea of six degrees of separation to get to any person on the planet, I came up with the idea to use a word cloud that would represent the top N words in all documents.

When you click on a word, (say "alpha") the resulting word cloud would represent the top N words for all the documents with "alpha" in it.

As you click, bravo -> charlie, etc. the list of documents gets smaller and smaller, until just your required document remains.

This has several advantages, you don't need to distinguish between words and numbers or need to "understand" the meaning of a word or interpret the user intent.

More importantly, the user doesn't need to know the relevant words or vocabulary, since they're all represented in the UI.

Enhancements include allowing for negative words, as-in, exclude documents with this word.

[–] [email protected] 1 points 4 months ago (1 children)

Ah that sounds really interesting! Does it scale OK? I guess you could index at a word level and filter quite quickly for quick searches, but it seems you're going to have to store the full text of every website?

[–] [email protected] 2 points 4 months ago (1 children)

You store just the word count for each word on each URL.

The search is pretty trivial in database terms since you don't need to do any wildcard or like matching.

[–] [email protected] 1 points 4 months ago (1 children)

Ah of course!

I guess one of the things the Google originally solved was that the internet if full of crap and not all sites should have equal weighing. With AI spam sites these days, you'd probably also need a method of weighting results?

[–] [email protected] 2 points 4 months ago (1 children)

We never got that far to test that kind of issue and while I've been reimplementing it locally to search through employment advertising, I'm not at a point where I'd be able to test such a thing.

The original implementation used a data store written by another team member and it made the original project much too complicated.

Today I'd likely use duckdb to implement it. My local version uses text files for a proof of concept implementation.

[–] [email protected] 1 points 4 months ago

It sounds like a really cool project regardless!