• 6 Posts
  • 9 Comments
Joined 2 months ago
cake
Cake day: February 17th, 2025

help-circle
  • Cattail@lemmy.worldOPtoFediverse@lemmy.worldFirst draft woes
    link
    fedilink
    English
    arrow-up
    1
    ·
    38 minutes ago

    I did make a classier for for videos that inputs title, tags, description, and closed caption into an LLM. I got roughly 1000 entries classified that way, issue is that most of them were non-english videos and then new videos come from somewhere on peertube that don’t hace these classifiers.

    Video processing is cool just expensive computationally. Also watchers could classify the videos themselves then use a cosine similarity (or whatever algo) on that. I did suggest to peertube to share the categories people say a video is with other people (like it’s a Mastodon post) eventually it morphed into an idea light weight peertube instance that only does api.










  • be better to store the video vector on an instance so that watchers can retrieve, just logistics. video vector (element) can be calculated anywhere just communicated to an instance, the idea is to be flexible. activityhub protocol has made the decisions easy the video vector has to be a .json element in a video json data.

    it would be better to store the results of a calculation to avoid repeated calculations. im looking into music classifications, and like the entire video can be sent to parse to see if its music or not, the tempo, genra, id assume that would be fairly costly to calculate or instance can send the video vector that states all that information



  • i havent made anything yet i just wanted to articulate that a basic algorythm can be done ethically where either instance/watcher/fediverse in general can make a vector to define a video and that could be shared via activity hub and the user can have a vector for themselves and even their own algorithm to sift through videos.

    im just starting and right now i have to figure out how to format the video vector do i want .json .csv .xml