package handler import ( "bytes" "context" "encoding/json" "fmt" "io" "net/http" "net/url" "regexp" "slices" "strings" "time" "github.com/PuerkitoBio/goquery" "github.com/felipemarinho97/torrent-indexer/magnet" "github.com/felipemarinho97/torrent-indexer/schema" goscrape "github.com/felipemarinho97/torrent-indexer/scrape" "github.com/felipemarinho97/torrent-indexer/utils" "github.com/hbollon/go-edlib" ) var comando = IndexerMeta{ URL: "https://comando.la/", SearchURL: "?s=", } var replacer = strings.NewReplacer( "janeiro", "01", "fevereiro", "02", "março", "03", "abril", "04", "maio", "05", "junho", "06", "julho", "07", "agosto", "08", "setembro", "09", "outubro", "10", "novembro", "11", "dezembro", "12", ) func (i *Indexer) HandlerComandoIndexer(w http.ResponseWriter, r *http.Request) { start := time.Now() defer func() { i.metrics.IndexerDuration.WithLabelValues("comando").Observe(time.Since(start).Seconds()) i.metrics.IndexerRequests.WithLabelValues("comando").Inc() }() ctx := r.Context() // supported query params: q, season, episode, page, filter_results q := r.URL.Query().Get("q") page := r.URL.Query().Get("page") // URL encode query param q = url.QueryEscape(q) url := comando.URL if q != "" { url = fmt.Sprintf("%s%s%s", url, comando.SearchURL, q) } else if page != "" { url = fmt.Sprintf("%spage/%s", url, page) } fmt.Println("URL:>", url) resp, err := i.requester.GetDocument(ctx, url) if err != nil { w.WriteHeader(http.StatusInternalServerError) err = json.NewEncoder(w).Encode(map[string]string{"error": err.Error()}) if err != nil { fmt.Println(err) } i.metrics.IndexerErrors.WithLabelValues("comando").Inc() return } defer resp.Close() doc, err := goquery.NewDocumentFromReader(resp) if err != nil { w.WriteHeader(http.StatusInternalServerError) err = json.NewEncoder(w).Encode(map[string]string{"error": err.Error()}) if err != nil { fmt.Println(err) } i.metrics.IndexerErrors.WithLabelValues("comando").Inc() return } var links []string doc.Find("article").Each(func(i int, s *goquery.Selection) { // get link from h2.entry-title > a link, _ := s.Find("h2.entry-title > a").Attr("href") links = append(links, link) }) var itChan = make(chan []schema.IndexedTorrent) var errChan = make(chan error) indexedTorrents := []schema.IndexedTorrent{} for _, link := range links { go func(link string) { torrents, err := getTorrents(ctx, i, link) if err != nil { fmt.Println(err) errChan <- err } itChan <- torrents }(link) } for i := 0; i < len(links); i++ { select { case torrents := <-itChan: indexedTorrents = append(indexedTorrents, torrents...) case err := <-errChan: fmt.Println(err) } } for i, it := range indexedTorrents { jLower := strings.ReplaceAll(strings.ToLower(fmt.Sprintf("%s %s", it.Title, it.OriginalTitle)), ".", " ") qLower := strings.ToLower(q) splitLength := 2 indexedTorrents[i].Similarity = edlib.JaccardSimilarity(jLower, qLower, splitLength) } // remove the ones with zero similarity if len(indexedTorrents) > 20 && r.URL.Query().Get("filter_results") != "" && r.URL.Query().Get("q") != "" { indexedTorrents = utils.Filter(indexedTorrents, func(it schema.IndexedTorrent) bool { return it.Similarity > 0 }) } // sort by similarity slices.SortFunc(indexedTorrents, func(i, j schema.IndexedTorrent) int { return int((j.Similarity - i.Similarity) * 1000) }) // send to search index go func() { _ = i.search.IndexTorrents(indexedTorrents) }() w.Header().Set("Content-Type", "application/json") err = json.NewEncoder(w).Encode(Response{ Results: indexedTorrents, Count: len(indexedTorrents), }) if err != nil { fmt.Println(err) } } func getTorrents(ctx context.Context, i *Indexer, link string) ([]schema.IndexedTorrent, error) { var indexedTorrents []schema.IndexedTorrent doc, err := getDocument(ctx, i, link) if err != nil { return nil, err } article := doc.Find("article") title := strings.Replace(article.Find(".entry-title").Text(), " - Download", "", -1) textContent := article.Find("div.entry-content") // div itemprop="datePublished" datePublished := strings.TrimSpace(article.Find("div[itemprop=\"datePublished\"]").Text()) // pattern: 10 de setembro de 2021 re := regexp.MustCompile(`(\d{2}) de (\w+) de (\d{4})`) matches := re.FindStringSubmatch(datePublished) var date time.Time if len(matches) > 0 { day := matches[1] month := matches[2] year := matches[3] datePublished = fmt.Sprintf("%s-%s-%s", year, replacer.Replace(month), day) date, err = time.Parse("2006-01-02", datePublished) if err != nil { return nil, err } } magnets := textContent.Find("a[href^=\"magnet\"]") var magnetLinks []string magnets.Each(func(i int, s *goquery.Selection) { magnetLink, _ := s.Attr("href") magnetLinks = append(magnetLinks, magnetLink) }) var audio []schema.Audio var year string var size []string article.Find("div.entry-content > p").Each(func(i int, s *goquery.Selection) { // pattern: // Título Traduzido: Fundação // Título Original: Foundation // IMDb: 7,5 // Ano de Lançamento: 2023 // Gênero: Ação | Aventura | Ficção // Formato: MKV // Qualidade: WEB-DL // Áudio: Português | Inglês // Idioma: Português | Inglês // Legenda: Português // Tamanho: – // Qualidade de Áudio: 10 // Qualidade de Vídeo: 10 // Duração: 59 Min. // Servidor: Torrent text := s.Text() audio = append(audio, findAudioFromText(text)...) y := findYearFromText(text, title) if y != "" { year = y } size = append(size, findSizesFromText(text)...) }) // find any link from imdb imdbLink := "" article.Find("a").Each(func(i int, s *goquery.Selection) { link, _ := s.Attr("href") re := regexp.MustCompile(`https://www.imdb.com/title/(tt\d+)`) matches := re.FindStringSubmatch(link) if len(matches) > 0 { imdbLink = matches[0] } }) size = stableUniq(size) var chanIndexedTorrent = make(chan schema.IndexedTorrent) // for each magnet link, create a new indexed torrent for it, magnetLink := range magnetLinks { it := it go func(it int, magnetLink string) { magnet, err := magnet.ParseMagnetUri(magnetLink) if err != nil { fmt.Println(err) } releaseTitle := magnet.DisplayName infoHash := magnet.InfoHash.String() trackers := magnet.Trackers magnetAudio := []schema.Audio{} if strings.Contains(strings.ToLower(releaseTitle), "dual") || strings.Contains(strings.ToLower(releaseTitle), "dublado") { magnetAudio = append(magnetAudio, audio...) } else if len(audio) > 1 { // remove portuguese audio, and append to magnetAudio for _, a := range audio { if a != schema.AudioPortuguese { magnetAudio = append(magnetAudio, a) } } } else { magnetAudio = append(magnetAudio, audio...) } peer, seed, err := goscrape.GetLeechsAndSeeds(ctx, i.redis, i.metrics, infoHash, trackers) if err != nil { fmt.Println(err) } title := processTitle(title, magnetAudio) // if the number of sizes is equal to the number of magnets, then assign the size to each indexed torrent in order var mySize string if len(size) == len(magnetLinks) { mySize = size[it] } ixt := schema.IndexedTorrent{ Title: appendAudioISO639_2Code(releaseTitle, magnetAudio), OriginalTitle: title, Details: link, Year: year, IMDB: imdbLink, Audio: magnetAudio, MagnetLink: magnetLink, Date: date, InfoHash: infoHash, Trackers: trackers, LeechCount: peer, SeedCount: seed, Size: mySize, } chanIndexedTorrent <- ixt }(it, magnetLink) } for i := 0; i < len(magnetLinks); i++ { it := <-chanIndexedTorrent indexedTorrents = append(indexedTorrents, it) } return indexedTorrents, nil } func stableUniq(s []string) []string { var uniq []map[string]interface{} m := make(map[string]map[string]interface{}) for i, v := range s { m[v] = map[string]interface{}{ "v": v, "i": i, } } // to order by index for _, v := range m { uniq = append(uniq, v) } // sort by index for i := 0; i < len(uniq); i++ { for j := i + 1; j < len(uniq); j++ { if uniq[i]["i"].(int) > uniq[j]["i"].(int) { uniq[i], uniq[j] = uniq[j], uniq[i] } } } // get only values var uniqValues []string for _, v := range uniq { uniqValues = append(uniqValues, v["v"].(string)) } return uniqValues } func findYearFromText(text string, title string) (year string) { re := regexp.MustCompile(`Lançamento: (.*)`) yearMatch := re.FindStringSubmatch(text) if len(yearMatch) > 0 { year = yearMatch[1] } if year == "" { re = regexp.MustCompile(`\((\d{4})\)`) yearMatch := re.FindStringSubmatch(title) if len(yearMatch) > 0 { year = yearMatch[1] } } return year } func findAudioFromText(text string) []schema.Audio { var audio []schema.Audio re := regexp.MustCompile(`(.udio|Idioma):.?(.*)`) audioMatch := re.FindStringSubmatch(text) if len(audioMatch) > 0 { sep := getSeparator(audioMatch[2]) langs_raw := strings.Split(audioMatch[2], sep) for _, lang := range langs_raw { lang = strings.TrimSpace(lang) a := schema.GetAudioFromString(lang) if a != nil { audio = append(audio, *a) } else { fmt.Println("unknown language:", lang) } } } return audio } func findSizesFromText(text string) []string { var sizes []string // everything that ends with GB or MB, using ',' or '.' as decimal separator re := regexp.MustCompile(`(\d+[\.,]?\d+) ?(GB|MB)`) sizesMatch := re.FindAllStringSubmatch(text, -1) if len(sizesMatch) > 0 { for _, size := range sizesMatch { sizes = append(sizes, size[0]) } } return sizes } func processTitle(title string, a []schema.Audio) string { // remove ' - Donwload' from title title = strings.Replace(title, " – Download", "", -1) // remove 'comando.la' from title title = strings.Replace(title, "comando.la", "", -1) // add audio ISO 639-2 code to title between () title = appendAudioISO639_2Code(title, a) return title } func appendAudioISO639_2Code(title string, a []schema.Audio) string { if len(a) > 0 { audio := []string{} for _, lang := range a { audio = append(audio, lang.String()) } title = fmt.Sprintf("%s (%s)", title, strings.Join(audio, ", ")) } return title } func getSeparator(s string) string { if strings.Contains(s, "|") { return "|" } else if strings.Contains(s, ",") { return "," } return " " } func getDocument(ctx context.Context, i *Indexer, link string) (*goquery.Document, error) { // try to get from redis first docCache, err := i.redis.Get(ctx, link) if err == nil { i.metrics.CacheHits.WithLabelValues("document_body").Inc() fmt.Printf("returning from long-lived cache: %s\n", link) return goquery.NewDocumentFromReader(io.NopCloser(bytes.NewReader(docCache))) } defer i.metrics.CacheMisses.WithLabelValues("document_body").Inc() resp, err := i.requester.GetDocument(ctx, link) if err != nil { return nil, err } defer resp.Close() body, err := io.ReadAll(resp) if err != nil { return nil, err } // set cache err = i.redis.Set(ctx, link, body) if err != nil { fmt.Println(err) } doc, err := goquery.NewDocumentFromReader(io.NopCloser(bytes.NewReader(body))) if err != nil { return nil, err } return doc, nil }