2024-12-29 00:56:47 +01:00
|
|
|
package main
|
|
|
|
|
|
|
|
|
|
import (
|
2025-01-04 16:53:46 +01:00
|
|
|
"crowsnest/cmd/crawler/extractors"
|
2025-01-06 19:58:01 +01:00
|
|
|
"crowsnest/internal/model/database"
|
2025-01-03 01:00:06 +01:00
|
|
|
"database/sql"
|
|
|
|
|
"log"
|
2025-01-07 09:32:57 +01:00
|
|
|
"os"
|
|
|
|
|
"time"
|
2024-12-29 00:56:47 +01:00
|
|
|
|
2025-01-07 09:32:57 +01:00
|
|
|
_ "github.com/lib/pq"
|
2024-12-29 00:56:47 +01:00
|
|
|
)
|
|
|
|
|
|
|
|
|
|
func main() {
|
2025-01-07 09:32:57 +01:00
|
|
|
// collect environement variables
|
|
|
|
|
databaseURL := os.Getenv("DB_URL")
|
|
|
|
|
|
|
|
|
|
// connect to database
|
|
|
|
|
db, err := sql.Open("postgres", databaseURL)
|
2025-01-04 16:53:46 +01:00
|
|
|
if err != nil {
|
|
|
|
|
log.Fatal(err)
|
|
|
|
|
}
|
2025-01-07 09:32:57 +01:00
|
|
|
defer db.Close()
|
2024-12-29 00:56:47 +01:00
|
|
|
|
2025-01-04 16:53:46 +01:00
|
|
|
// collect websites
|
2025-01-07 09:32:57 +01:00
|
|
|
//coll := collectors.Collector{
|
|
|
|
|
// Responses: &database.ResponseModel{DB: db},
|
|
|
|
|
//}
|
2025-01-03 01:00:06 +01:00
|
|
|
|
2025-01-04 16:53:46 +01:00
|
|
|
//coll.Spiegel()
|
|
|
|
|
//coll.Zeit()
|
2024-12-29 00:56:47 +01:00
|
|
|
|
2025-01-04 16:53:46 +01:00
|
|
|
// extract articles from websites
|
|
|
|
|
extr := extractors.Extractor{
|
2025-01-06 19:58:01 +01:00
|
|
|
Responses: &database.ResponseModel{DB: db},
|
|
|
|
|
Articles: &database.ArticleModel{DB: db},
|
2024-12-29 00:56:47 +01:00
|
|
|
}
|
2025-01-04 16:53:46 +01:00
|
|
|
|
2025-01-07 09:32:57 +01:00
|
|
|
for {
|
|
|
|
|
extr.Spiegel()
|
|
|
|
|
time.Sleep(5 * time.Second)
|
|
|
|
|
}
|
2024-12-29 00:56:47 +01:00
|
|
|
}
|