Skip to content

Commit

Permalink
align classpath config and file config
Browse files Browse the repository at this point in the history
  • Loading branch information
codingPF committed Nov 8, 2023
1 parent 402a0b8 commit 16e2afa
Show file tree
Hide file tree
Showing 2 changed files with 60 additions and 16 deletions.
2 changes: 1 addition & 1 deletion MServer-Config.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -182,7 +182,7 @@ copySettings:
# En- / disables FTP
copyEnabled: false

# The paths where to safe the film list files.SrfTopicOverviewTask
# The paths where to safe the film list files.
# WARNING: You can only set the path for film list formats you listed in "filmlistSaveFormats".
# Required if enabled
copyTargetFilePaths:
Expand Down
74 changes: 59 additions & 15 deletions src/main/resources/MServer-Config.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -7,6 +7,9 @@ maximumCpuThreads: 16
# If set to 0 the server runs without a time limit.
maximumServerDurationInMinutes: 0

# Rate limiter
maximumRequestsPerSecond: 999.0

# These Sender will NOT be crawled.
# If no Sender are included the server will crawl all Sender but these.
#senderExcluded:
Expand Down Expand Up @@ -77,22 +80,40 @@ writeFilmlistIdFileEnabled: true
#The fimlist id file path
filmlistIdFilePath: target/filmlist.id

# Sets if a filmlist should be imported
filmlistImporEnabled: false

# The format of the film list to import.
# Possible are: JSON, OLD_JSON, JSON_COMPRESSED_XZ, OLD_JSON_COMPRESSED_XZ, JSON_COMPRESSED_GZIP, OLD_JSON_COMPRESSED_BZIP, JSON_COMPRESSED_GZIP, OLD_JSON_COMPRESSED_BZIP
#filmlistImportFormat: OLD_JSON_COMPRESSED_XZ

# The path/URL of the film list to import.
#filmlistImportLocation: http://verteiler1.mediathekview.de/Filmliste-akt.xz
# import additional filmlist sources
importFilmlistConfigurations :
- active: false
path: "someCrawlerlist.json"
format: OLD_JSON
createDiff: false
checkImportListUrl: false
- active: false
path: "someMoreCrawlerlist.json"
format: OLD_JSON
createDiff: false
checkImportListUrl: false
- active: false
path: "https://verteiler1.mediathekview.de/filme-org.xz"
format: OLD_JSON_COMPRESSED_XZ
createDiff: true
checkImportListUrl: true

# film url is consider invalid if the size is below the minSize
checkImportListUrlMinSize: 5012

# abort url checking after x sec
checkImportListUrlTimeoutInSec: 1800

#### Default crawler configurations ####
# The maximum amount of URLs to be processed per task.
maximumUrlsPerTask: 50

# The maximum duration in minutes a crawler may run.
maximumCrawlDurationInMinutes: 60
maximumCrawlDurationInMinutes: 120

# Enables the topics search
# maximumSubpages limits the depth of the topics search
topicsSearchEnabled: false

# The maximum amount of sub pages to be crawled.<br>
# Example: If a Sendung overview side has 10 pages with videos for this Sendung and
Expand All @@ -113,9 +134,20 @@ socketTimeoutInSeconds: 60
senderConfigurations:
ARD:
# Actually the ARD has a maximum of 6 days in the past
maximumDaysForSendungVerpasstSection: 6
maximumDaysForSendungVerpasstSection: 1
#2,4,8 ok
maximumUrlsPerTask: 32
#10,20,40 ok
maximumSubpages: 0
ORF:
#2,4,8 ok
maximumUrlsPerTask: 40
ARTE_DE:
maximumDaysForSendungVerpasstSectionFuture: 21
maximumUrlsPerTask: 1
maximumDaysForSendungVerpasstSectionFuture: 0
maximumRequestsPerSecond: 2.0
ARTE_FR:
maximumDaysForSendungVerpasstSectionFuture: 0
# The maximum amount of URLs to be processed per task.
# maximumUrlsPerTask: 25
# The maximum duration in minutes a crawler may run.
Expand All @@ -125,9 +157,21 @@ senderConfigurations:
# the amount set by this is 5 then the crawler crawls pages 1 to 5.
# maximumSubpages: 3
KIKA:
socketTimeoutInSeconds: 120


maximumSubpages: 2
maximumRequestsPerSecond: 8.0
SR:
maximumRequestsPerSecond: 2.0
ZDF:
maximumRequestsPerSecond: 10.0
FUNK:
maximumUrlsPerTask: 99
DW:
maximumSubpages: 0

# configure string variables
crawlerApiParams:
FUNK_REQUEST_TOKEN: 137782e774d7cadc93dcbffbbde0ce9c

#### COPY ####
copySettings:
# En- / disables FTP
Expand Down

0 comments on commit 16e2afa

Please sign in to comment.