mirror of
https://github.com/TheAnachronism/docspell.git
synced 2025-04-04 10:29:34 +00:00
parent
bc9a483143
commit
81d8b6c9c1
@ -11,6 +11,7 @@ import cats.implicits._
|
|||||||
|
|
||||||
import docspell.backend.signup.{Config => SignupConfig}
|
import docspell.backend.signup.{Config => SignupConfig}
|
||||||
import docspell.common._
|
import docspell.common._
|
||||||
|
import docspell.common.syntax.string._
|
||||||
import docspell.store.file.FileRepositoryConfig
|
import docspell.store.file.FileRepositoryConfig
|
||||||
import docspell.store.{JdbcConfig, SchemaMigrateConfig}
|
import docspell.store.{JdbcConfig, SchemaMigrateConfig}
|
||||||
|
|
||||||
@ -51,6 +52,12 @@ object Config {
|
|||||||
stores.get(id).map(FileRepositoryConfig.fromFileStoreConfig(chunkSize, _))
|
stores.get(id).map(FileRepositoryConfig.fromFileStoreConfig(chunkSize, _))
|
||||||
|
|
||||||
def validate: ValidatedNec[String, Files] = {
|
def validate: ValidatedNec[String, Files] = {
|
||||||
|
def filterEmptyRegion =
|
||||||
|
copy(stores = stores.view.mapValues {
|
||||||
|
case c: FileStoreConfig.S3 => c.copy(region = c.region.flatMap(_.asNonBlank))
|
||||||
|
case c => c
|
||||||
|
}.toMap)
|
||||||
|
|
||||||
val storesEmpty =
|
val storesEmpty =
|
||||||
if (enabledStores.isEmpty)
|
if (enabledStores.isEmpty)
|
||||||
Validated.invalidNec(
|
Validated.invalidNec(
|
||||||
@ -65,7 +72,7 @@ object Config {
|
|||||||
Validated.invalidNec(s"Default file store not present: ${defaultStore.id}")
|
Validated.invalidNec(s"Default file store not present: ${defaultStore.id}")
|
||||||
}
|
}
|
||||||
|
|
||||||
(storesEmpty |+| defaultStorePresent).map(_ => this)
|
(storesEmpty |+| defaultStorePresent).map(_ => filterEmptyRegion)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -29,11 +29,12 @@ object FileStoreConfig {
|
|||||||
endpoint: String,
|
endpoint: String,
|
||||||
accessKey: String,
|
accessKey: String,
|
||||||
secretKey: String,
|
secretKey: String,
|
||||||
|
region: Option[String],
|
||||||
bucket: String
|
bucket: String
|
||||||
) extends FileStoreConfig {
|
) extends FileStoreConfig {
|
||||||
val storeType = FileStoreType.S3
|
val storeType = FileStoreType.S3
|
||||||
|
|
||||||
override def toString =
|
override def toString =
|
||||||
s"S3(enabled=$enabled, endpoint=$endpoint, bucket=$bucket, accessKey=$accessKey, secretKey=***)"
|
s"S3(enabled=$enabled, endpoint=$endpoint, bucket=$bucket, accessKey=$accessKey, secretKey=***, region=$region)"
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -739,6 +739,7 @@ Docpell Update Check
|
|||||||
access-key = "username"
|
access-key = "username"
|
||||||
secret-key = "password"
|
secret-key = "password"
|
||||||
bucket = "docspell"
|
bucket = "docspell"
|
||||||
|
region = ""
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -60,7 +60,7 @@ object ConfigFile {
|
|||||||
cfg => cfg.updateCheck.enabled && cfg.updateCheck.subject.els.isEmpty,
|
cfg => cfg.updateCheck.enabled && cfg.updateCheck.subject.els.isEmpty,
|
||||||
"No subject given for enabled update check!"
|
"No subject given for enabled update check!"
|
||||||
),
|
),
|
||||||
Validation(cfg => cfg.files.validate.map(_ => cfg)),
|
Validation(cfg => cfg.files.validate.map(f => cfg.copy(files = f))),
|
||||||
Validation.failWhen(
|
Validation.failWhen(
|
||||||
cfg =>
|
cfg =>
|
||||||
cfg.fullTextSearch.enabled &&
|
cfg.fullTextSearch.enabled &&
|
||||||
|
@ -493,6 +493,7 @@ docspell.server {
|
|||||||
access-key = "username"
|
access-key = "username"
|
||||||
secret-key = "password"
|
secret-key = "password"
|
||||||
bucket = "docspell"
|
bucket = "docspell"
|
||||||
|
region = ""
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -113,7 +113,9 @@ object ConfigFile {
|
|||||||
}
|
}
|
||||||
|
|
||||||
def filesValidate: Validation[Config] =
|
def filesValidate: Validation[Config] =
|
||||||
Validation(cfg => cfg.backend.files.validate.map(_ => cfg))
|
Validation(cfg =>
|
||||||
|
cfg.backend.files.validate.map(f => cfg.copy(backend = cfg.backend.copy(files = f)))
|
||||||
|
)
|
||||||
|
|
||||||
def postgresFtsValidate: Validation[Config] =
|
def postgresFtsValidate: Validation[Config] =
|
||||||
Validation.failWhen(
|
Validation.failWhen(
|
||||||
|
@ -108,11 +108,22 @@ object BinnyUtils {
|
|||||||
JdbcStoreConfig("filechunk", chunkSize, BinnyUtils.TikaContentTypeDetect)
|
JdbcStoreConfig("filechunk", chunkSize, BinnyUtils.TikaContentTypeDetect)
|
||||||
GenericJdbcStore[F](ds, LoggerAdapter(logger), jdbcConfig)
|
GenericJdbcStore[F](ds, LoggerAdapter(logger), jdbcConfig)
|
||||||
|
|
||||||
case FileRepositoryConfig.S3(endpoint, accessKey, secretKey, bucket, chunkSize) =>
|
case FileRepositoryConfig.S3(
|
||||||
|
endpoint,
|
||||||
|
accessKey,
|
||||||
|
secretKey,
|
||||||
|
bucket,
|
||||||
|
region,
|
||||||
|
chunkSize
|
||||||
|
) =>
|
||||||
val keyMapping = S3KeyMapping.constant(bucket)
|
val keyMapping = S3KeyMapping.constant(bucket)
|
||||||
val minioCfg = MinioConfig
|
val minioCfg = MinioConfig
|
||||||
.default(endpoint, accessKey, secretKey, keyMapping)
|
.default(endpoint, accessKey, secretKey, keyMapping)
|
||||||
.copy(chunkSize = chunkSize, detect = BinnyUtils.TikaContentTypeDetect)
|
.copy(
|
||||||
|
chunkSize = chunkSize,
|
||||||
|
detect = BinnyUtils.TikaContentTypeDetect,
|
||||||
|
region = region
|
||||||
|
)
|
||||||
|
|
||||||
MinioBinaryStore[F](minioCfg, LoggerAdapter(logger))
|
MinioBinaryStore[F](minioCfg, LoggerAdapter(logger))
|
||||||
|
|
||||||
|
@ -10,7 +10,7 @@ import fs2.io.file.Path
|
|||||||
|
|
||||||
import docspell.common.FileStoreConfig
|
import docspell.common.FileStoreConfig
|
||||||
|
|
||||||
sealed trait FileRepositoryConfig {}
|
sealed trait FileRepositoryConfig
|
||||||
|
|
||||||
object FileRepositoryConfig {
|
object FileRepositoryConfig {
|
||||||
|
|
||||||
@ -21,6 +21,7 @@ object FileRepositoryConfig {
|
|||||||
accessKey: String,
|
accessKey: String,
|
||||||
secretKey: String,
|
secretKey: String,
|
||||||
bucketName: String,
|
bucketName: String,
|
||||||
|
region: Option[String],
|
||||||
chunkSize: Int
|
chunkSize: Int
|
||||||
) extends FileRepositoryConfig
|
) extends FileRepositoryConfig
|
||||||
|
|
||||||
@ -30,8 +31,15 @@ object FileRepositoryConfig {
|
|||||||
cfg match {
|
cfg match {
|
||||||
case FileStoreConfig.DefaultDatabase(_) =>
|
case FileStoreConfig.DefaultDatabase(_) =>
|
||||||
FileRepositoryConfig.Database(chunkSize)
|
FileRepositoryConfig.Database(chunkSize)
|
||||||
case FileStoreConfig.S3(_, endpoint, accessKey, secretKey, bucket) =>
|
case FileStoreConfig.S3(_, endpoint, accessKey, secretKey, region, bucket) =>
|
||||||
FileRepositoryConfig.S3(endpoint, accessKey, secretKey, bucket, chunkSize)
|
FileRepositoryConfig.S3(
|
||||||
|
endpoint,
|
||||||
|
accessKey,
|
||||||
|
secretKey,
|
||||||
|
bucket,
|
||||||
|
region.map(_.trim).filter(_.nonEmpty),
|
||||||
|
chunkSize
|
||||||
|
)
|
||||||
case FileStoreConfig.FileSystem(_, directory) =>
|
case FileStoreConfig.FileSystem(_, directory) =>
|
||||||
FileRepositoryConfig.Directory(directory, chunkSize)
|
FileRepositoryConfig.Directory(directory, chunkSize)
|
||||||
}
|
}
|
||||||
|
@ -73,6 +73,7 @@ files {
|
|||||||
access-key = "username"
|
access-key = "username"
|
||||||
secret-key = "password"
|
secret-key = "password"
|
||||||
bucket = "docspell"
|
bucket = "docspell"
|
||||||
|
# region = "…"
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -54,8 +54,8 @@ For linuxes, the default location is `~/.config/dsc/config.toml`. You
|
|||||||
can give a config file explicitly via an option or the environment
|
can give a config file explicitly via an option or the environment
|
||||||
variable `DSC_CONFIG`.
|
variable `DSC_CONFIG`.
|
||||||
|
|
||||||
If you use the [pass](https://passwordstore.org) password manager, you
|
If you use the [pass](https://www.passwordstore.org) password manager,
|
||||||
can add your password entry to the config file as well.
|
you can add your password entry to the config file as well.
|
||||||
|
|
||||||
## Output format
|
## Output format
|
||||||
|
|
||||||
|
Loading…
x
Reference in New Issue
Block a user