Merge branch 'release/v2.5.0' into develop
This commit is contained in:
commit
80b0d3ffd6
@ -1,3 +0,0 @@
|
|||||||
{
|
|
||||||
"kafka": {}
|
|
||||||
}
|
|
@ -1,5 +0,0 @@
|
|||||||
{
|
|
||||||
"Zookeeper": {
|
|
||||||
|
|
||||||
}
|
|
||||||
}
|
|
@ -1,6 +0,0 @@
|
|||||||
{
|
|
||||||
"Zookeeper": {
|
|
||||||
"host": "localhost",
|
|
||||||
"port": 2181
|
|
||||||
}
|
|
||||||
}
|
|
@ -7,5 +7,6 @@ debian/build/buildtools/config/nginx/onlyoffice*.conf etc
|
|||||||
debian/build/buildtools/install/docker/config/nginx/letsencrypt* etc/openresty/includes
|
debian/build/buildtools/install/docker/config/nginx/letsencrypt* etc/openresty/includes
|
||||||
debian/build/buildtools/config/nginx/includes/onlyoffice*.conf etc/openresty/includes
|
debian/build/buildtools/config/nginx/includes/onlyoffice*.conf etc/openresty/includes
|
||||||
debian/build/publish/web/public/* var/www/{{product}}/public
|
debian/build/publish/web/public/* var/www/{{product}}/public
|
||||||
|
debian/build/campaigns/src/campaigns/* var/www/{{product}}/public/campaigns
|
||||||
debian/build/publish/web/client/* var/www/{{product}}/client
|
debian/build/publish/web/client/* var/www/{{product}}/client
|
||||||
debian/build/publish/web/management/* var/www/{{product}}/management
|
debian/build/publish/web/management/* var/www/{{product}}/management
|
||||||
|
@ -46,6 +46,7 @@ check_archives:
|
|||||||
@$(call extract_archive,${SOURCE_PATH}/client.tar.gz,client,-C ${BUILD_PATH})
|
@$(call extract_archive,${SOURCE_PATH}/client.tar.gz,client,-C ${BUILD_PATH})
|
||||||
@$(call extract_archive,${SOURCE_PATH}/dictionaries.tar.gz,dictionaries,-C ${CLENT_PATH}/common/Tests/Frontend.Translations.Tests)
|
@$(call extract_archive,${SOURCE_PATH}/dictionaries.tar.gz,dictionaries,-C ${CLENT_PATH}/common/Tests/Frontend.Translations.Tests)
|
||||||
@$(call extract_archive,${SOURCE_PATH}/DocStore.tar.gz,DocStore,-C ${SERVER_PATH}/products/ASC.Files/Server)
|
@$(call extract_archive,${SOURCE_PATH}/DocStore.tar.gz,DocStore,-C ${SERVER_PATH}/products/ASC.Files/Server)
|
||||||
|
@$(call extract_archive,${SOURCE_PATH}/campaigns.tar.gz,campaigns,-C ${BUILD_PATH})
|
||||||
@echo "Source archives check passed."
|
@echo "Source archives check passed."
|
||||||
|
|
||||||
override_dh_auto_build: check_archives
|
override_dh_auto_build: check_archives
|
||||||
|
@ -15,21 +15,6 @@
|
|||||||
APP_DOTNET_ENV=""
|
APP_DOTNET_ENV=""
|
||||||
EXTERNAL_PORT="80"
|
EXTERNAL_PORT="80"
|
||||||
|
|
||||||
# zookeeper #
|
|
||||||
ZOO_PORT=2181
|
|
||||||
ZOO_HOST=${CONTAINER_PREFIX}zookeeper
|
|
||||||
ZOO_SERVER=server.1=${ZOO_HOST}:2888:3888
|
|
||||||
|
|
||||||
# kafka #
|
|
||||||
KAFKA_HOST=${CONTAINER_PREFIX}kafka
|
|
||||||
KAFKA_ADVERTISED_LISTENERS=LISTENER_DOCKER_INTERNAL://${KAFKA_HOST}:9092
|
|
||||||
KAFKA_LISTENER_SECURITY_PROTOCOL_MAP=LISTENER_DOCKER_INTERNAL:PLAINTEXT,LISTENER_DOCKER_EXTERNAL:PLAINTEXT
|
|
||||||
KAFKA_INTER_BROKER_LISTENER_NAME=LISTENER_DOCKER_INTERNAL
|
|
||||||
KAFKA_ZOOKEEPER_CONNECT=${ZOO_HOST}:2181
|
|
||||||
KAFKA_BROKER_ID=1
|
|
||||||
KAFKA_LOG4J_LOGGERS=kafka.controller=INFO,kafka.producer.async.DefaultEventHandler=INFO,state.change.logger=INFO
|
|
||||||
KAFKA_OFFSETS_TOPIC_REPLICATION_FACTOR=1
|
|
||||||
|
|
||||||
# elasticsearch #
|
# elasticsearch #
|
||||||
ELK_CONTAINER_NAME=${CONTAINER_PREFIX}opensearch
|
ELK_CONTAINER_NAME=${CONTAINER_PREFIX}opensearch
|
||||||
ELK_SHEME=http
|
ELK_SHEME=http
|
||||||
|
@ -57,7 +57,6 @@ ELK_PORT = os.environ["ELK_PORT"] if environ.get("ELK_PORT") else "9200"
|
|||||||
ELK_THREADS = os.environ["ELK_THREADS"] if environ.get("ELK_THREADS") else "1"
|
ELK_THREADS = os.environ["ELK_THREADS"] if environ.get("ELK_THREADS") else "1"
|
||||||
ELK_CONNECTION_HOST = ELK_HOST if ELK_HOST else ELK_CONTAINER_NAME
|
ELK_CONNECTION_HOST = ELK_HOST if ELK_HOST else ELK_CONTAINER_NAME
|
||||||
|
|
||||||
KAFKA_HOST = os.environ["KAFKA_HOST"] if environ.get("KAFKA_HOST") else "kafka:9092"
|
|
||||||
RUN_FILE = sys.argv[1] if (len(sys.argv) > 1) else "none"
|
RUN_FILE = sys.argv[1] if (len(sys.argv) > 1) else "none"
|
||||||
LOG_FILE = sys.argv[2] if (len(sys.argv) > 2) else "none"
|
LOG_FILE = sys.argv[2] if (len(sys.argv) > 2) else "none"
|
||||||
CORE_EVENT_BUS = sys.argv[3] if (len(sys.argv) > 3) else ""
|
CORE_EVENT_BUS = sys.argv[3] if (len(sys.argv) > 3) else ""
|
||||||
@ -224,11 +223,6 @@ jsonData["elastic"]["Port"] = ELK_PORT
|
|||||||
jsonData["elastic"]["Threads"] = ELK_THREADS
|
jsonData["elastic"]["Threads"] = ELK_THREADS
|
||||||
writeJsonFile(filePath, jsonData)
|
writeJsonFile(filePath, jsonData)
|
||||||
|
|
||||||
filePath = "/app/onlyoffice/config/kafka.json"
|
|
||||||
jsonData = openJsonFile(filePath)
|
|
||||||
jsonData.update({"kafka": {"BootstrapServers": KAFKA_HOST}})
|
|
||||||
writeJsonFile(filePath, jsonData)
|
|
||||||
|
|
||||||
filePath = "/app/onlyoffice/config/socket.json"
|
filePath = "/app/onlyoffice/config/socket.json"
|
||||||
jsonData = openJsonFile(filePath)
|
jsonData = openJsonFile(filePath)
|
||||||
updateJsonData(jsonData,"$.socket.port", SERVICE_PORT)
|
updateJsonData(jsonData,"$.socket.port", SERVICE_PORT)
|
||||||
|
@ -32,7 +32,6 @@ x-service: &x-service-base
|
|||||||
DOCUMENT_SERVER_URL_PUBLIC: ${DOCUMENT_SERVER_URL_PUBLIC}
|
DOCUMENT_SERVER_URL_PUBLIC: ${DOCUMENT_SERVER_URL_PUBLIC}
|
||||||
DOCUMENT_CONTAINER_NAME: ${DOCUMENT_CONTAINER_NAME}
|
DOCUMENT_CONTAINER_NAME: ${DOCUMENT_CONTAINER_NAME}
|
||||||
DOCUMENT_SERVER_URL_EXTERNAL: ${DOCUMENT_SERVER_URL_EXTERNAL}
|
DOCUMENT_SERVER_URL_EXTERNAL: ${DOCUMENT_SERVER_URL_EXTERNAL}
|
||||||
KAFKA_HOST: ${KAFKA_HOST}
|
|
||||||
ELK_CONTAINER_NAME: ${ELK_CONTAINER_NAME}
|
ELK_CONTAINER_NAME: ${ELK_CONTAINER_NAME}
|
||||||
ELK_SHEME: ${ELK_SHEME}
|
ELK_SHEME: ${ELK_SHEME}
|
||||||
ELK_HOST: ${ELK_HOST}
|
ELK_HOST: ${ELK_HOST}
|
||||||
|
@ -1,42 +0,0 @@
|
|||||||
services:
|
|
||||||
onlyoffice-zookeeper:
|
|
||||||
image: zookeeper:latest
|
|
||||||
container_name: ${ZOO_HOST}
|
|
||||||
restart: always
|
|
||||||
expose:
|
|
||||||
- "2181"
|
|
||||||
environment:
|
|
||||||
ZOO_MY_ID: 1
|
|
||||||
ZOO_PORT: ${ZOO_PORT:-2181}
|
|
||||||
ZOO_SERVER: ${ZOO_SERVER}
|
|
||||||
volumes:
|
|
||||||
- /app/onlyoffice/data/zookeeper/zoo_data:/data
|
|
||||||
- /app/onlyoffice/data/zookeeper/zoo_log:/datalog
|
|
||||||
|
|
||||||
onlyoffice-kafka:
|
|
||||||
image: confluentinc/cp-kafka:latest
|
|
||||||
container_name: ${KAFKA_HOST}
|
|
||||||
restart: always
|
|
||||||
expose:
|
|
||||||
- "9092"
|
|
||||||
depends_on:
|
|
||||||
- onlyoffice-zookeeper
|
|
||||||
environment:
|
|
||||||
KAFKA_ADVERTISED_LISTENERS: ${KAFKA_ADVERTISED_LISTENERS}
|
|
||||||
KAFKA_LISTENER_SECURITY_PROTOCOL_MAP: ${KAFKA_LISTENER_SECURITY_PROTOCOL_MAP}
|
|
||||||
KAFKA_INTER_BROKER_LISTENER_NAME: ${KAFKA_INTER_BROKER_LISTENER_NAME}
|
|
||||||
KAFKA_ZOOKEEPER_CONNECT: ${KAFKA_ZOOKEEPER_CONNECT}
|
|
||||||
KAFKA_BROKER_ID: ${KAFKA_BROKER_ID}
|
|
||||||
KAFKA_LOG4J_LOGGERS: ${KAFKA_LOG4J_LOGGERS}
|
|
||||||
KAFKA_OFFSETS_TOPIC_REPLICATION_FACTOR: ${KAFKA_OFFSETS_TOPIC_REPLICATION_FACTOR}
|
|
||||||
volumes:
|
|
||||||
#- /app/onlyoffice/data/kafka_data:/var/lib/kafka/data
|
|
||||||
- kafka_data:/var/lib/kafka/data
|
|
||||||
|
|
||||||
volumes:
|
|
||||||
kafka_data:
|
|
||||||
|
|
||||||
networks:
|
|
||||||
default:
|
|
||||||
name: ${NETWORK_NAME}
|
|
||||||
external: true
|
|
@ -33,7 +33,6 @@ x-service:
|
|||||||
DOCUMENT_SERVER_URL_PUBLIC: ${DOCUMENT_SERVER_URL_PUBLIC}
|
DOCUMENT_SERVER_URL_PUBLIC: ${DOCUMENT_SERVER_URL_PUBLIC}
|
||||||
DOCUMENT_CONTAINER_NAME: ${DOCUMENT_CONTAINER_NAME}
|
DOCUMENT_CONTAINER_NAME: ${DOCUMENT_CONTAINER_NAME}
|
||||||
DOCUMENT_SERVER_URL_EXTERNAL: ${DOCUMENT_SERVER_URL_EXTERNAL}
|
DOCUMENT_SERVER_URL_EXTERNAL: ${DOCUMENT_SERVER_URL_EXTERNAL}
|
||||||
KAFKA_HOST: ${KAFKA_HOST}
|
|
||||||
ELK_CONTAINER_NAME: ${ELK_CONTAINER_NAME}
|
ELK_CONTAINER_NAME: ${ELK_CONTAINER_NAME}
|
||||||
ELK_SHEME: ${ELK_SHEME}
|
ELK_SHEME: ${ELK_SHEME}
|
||||||
ELK_HOST: ${ELK_HOST}
|
ELK_HOST: ${ELK_HOST}
|
||||||
|
@ -16,6 +16,7 @@ mkdir -p "%{buildroot}%{buildpath}/services/ASC.Data.Backup.BackgroundTasks/"
|
|||||||
mkdir -p "%{buildroot}%{buildpath}/services/ASC.ClearEvents/"
|
mkdir -p "%{buildroot}%{buildpath}/services/ASC.ClearEvents/"
|
||||||
mkdir -p "%{buildroot}%{buildpath}/services/ASC.ApiSystem/"
|
mkdir -p "%{buildroot}%{buildpath}/services/ASC.ApiSystem/"
|
||||||
mkdir -p "%{buildroot}%{buildpath}/public/"
|
mkdir -p "%{buildroot}%{buildpath}/public/"
|
||||||
|
mkdir -p "%{buildroot}%{buildpath}/public/campaigns/"
|
||||||
mkdir -p "%{buildroot}%{buildpath}/products/ASC.People/server/"
|
mkdir -p "%{buildroot}%{buildpath}/products/ASC.People/server/"
|
||||||
mkdir -p "%{buildroot}%{buildpath}/products/ASC.People/client/"
|
mkdir -p "%{buildroot}%{buildpath}/products/ASC.People/client/"
|
||||||
mkdir -p "%{buildroot}%{buildpath}/products/ASC.Login/login/"
|
mkdir -p "%{buildroot}%{buildpath}/products/ASC.Login/login/"
|
||||||
@ -37,6 +38,7 @@ mkdir -p "%{buildroot}%{_sysconfdir}/logrotate.d"
|
|||||||
mkdir -p "%{buildroot}%{_docdir}/%{name}-%{version}-%{release}/"
|
mkdir -p "%{buildroot}%{_docdir}/%{name}-%{version}-%{release}/"
|
||||||
mkdir -p "%{buildroot}%{_bindir}/"
|
mkdir -p "%{buildroot}%{_bindir}/"
|
||||||
cp -rf %{_builddir}/publish/web/public/* "%{buildroot}%{buildpath}/public/"
|
cp -rf %{_builddir}/publish/web/public/* "%{buildroot}%{buildpath}/public/"
|
||||||
|
cp -rf %{_builddir}/campaigns/src/campaigns/* "%{buildroot}%{buildpath}/public/campaigns"
|
||||||
cp -rf %{_builddir}/publish/web/login/* "%{buildroot}%{buildpath}/products/ASC.Login/login/"
|
cp -rf %{_builddir}/publish/web/login/* "%{buildroot}%{buildpath}/products/ASC.Login/login/"
|
||||||
cp -rf %{_builddir}/publish/web/editor/* "%{buildroot}%{buildpath}/products/ASC.Files/editor/"
|
cp -rf %{_builddir}/publish/web/editor/* "%{buildroot}%{buildpath}/products/ASC.Files/editor/"
|
||||||
cp -rf %{_builddir}/publish/web/editor/.next/* "%{buildroot}%{buildpath}/products/ASC.Files/editor/.next/"
|
cp -rf %{_builddir}/publish/web/editor/.next/* "%{buildroot}%{buildpath}/products/ASC.Files/editor/.next/"
|
||||||
|
@ -25,7 +25,8 @@ Source1: https://github.com/ONLYOFFICE/%{product}-client/archive/master.t
|
|||||||
Source2: https://github.com/ONLYOFFICE/%{product}-server/archive/master.tar.gz#/server.tar.gz
|
Source2: https://github.com/ONLYOFFICE/%{product}-server/archive/master.tar.gz#/server.tar.gz
|
||||||
Source3: https://github.com/ONLYOFFICE/document-templates/archive/main/community-server.tar.gz#/DocStore.tar.gz
|
Source3: https://github.com/ONLYOFFICE/document-templates/archive/main/community-server.tar.gz#/DocStore.tar.gz
|
||||||
Source4: https://github.com/ONLYOFFICE/dictionaries/archive/master.tar.gz#/dictionaries.tar.gz
|
Source4: https://github.com/ONLYOFFICE/dictionaries/archive/master.tar.gz#/dictionaries.tar.gz
|
||||||
Source5: %{product}.rpmlintrc
|
Source5: https://github.com/ONLYOFFICE/ASC.Web.Campaigns/archive/master.tar.gz#/campaigns.tar.gz
|
||||||
|
Source6: %{product}.rpmlintrc
|
||||||
|
|
||||||
BuildRequires: nodejs >= 18.0
|
BuildRequires: nodejs >= 18.0
|
||||||
BuildRequires: yarn
|
BuildRequires: yarn
|
||||||
@ -66,9 +67,10 @@ rm -rf %{_rpmdir}/%{_arch}/%{name}-* %{_builddir}/*
|
|||||||
tar -xf %{SOURCE0} --transform='s,^[^/]\+,buildtools,' -C %{_builddir}
|
tar -xf %{SOURCE0} --transform='s,^[^/]\+,buildtools,' -C %{_builddir}
|
||||||
tar -xf %{SOURCE1} --transform='s,^[^/]\+,client,' -C %{_builddir}
|
tar -xf %{SOURCE1} --transform='s,^[^/]\+,client,' -C %{_builddir}
|
||||||
tar -xf %{SOURCE2} --transform='s,^[^/]\+,server,' -C %{_builddir}
|
tar -xf %{SOURCE2} --transform='s,^[^/]\+,server,' -C %{_builddir}
|
||||||
tar -xf %{SOURCE4} --transform='s,^[^/]\+,dictionaries,' -C %{_builddir}/client/common/Tests/Frontend.Translations.Tests
|
|
||||||
tar -xf %{SOURCE3} --transform='s,^[^/]\+,DocStore,' -C %{_builddir}/server/products/ASC.Files/Server
|
tar -xf %{SOURCE3} --transform='s,^[^/]\+,DocStore,' -C %{_builddir}/server/products/ASC.Files/Server
|
||||||
cp %{SOURCE5} .
|
tar -xf %{SOURCE4} --transform='s,^[^/]\+,dictionaries,' -C %{_builddir}/client/common/Tests/Frontend.Translations.Tests
|
||||||
|
tar -xf %{SOURCE5} --transform='s,^[^/]\+,campaigns,' -C %{_builddir}
|
||||||
|
cp %{SOURCE6} .
|
||||||
|
|
||||||
%include build.spec
|
%include build.spec
|
||||||
|
|
||||||
|
@ -17,6 +17,7 @@ if defined SecondArg (
|
|||||||
)
|
)
|
||||||
|
|
||||||
xcopy "%PathToRepository%\publish\web\public" "%PathToAppFolder%\public" /s /y /b /i
|
xcopy "%PathToRepository%\publish\web\public" "%PathToAppFolder%\public" /s /y /b /i
|
||||||
|
xcopy "%PathToRepository%\campaigns\src\campaigns" "%PathToAppFolder%\public\campaigns" /s /y /b /i
|
||||||
xcopy "%PathToRepository%\publish\web\client" "%PathToAppFolder%\client" /s /y /b /i
|
xcopy "%PathToRepository%\publish\web\client" "%PathToAppFolder%\client" /s /y /b /i
|
||||||
xcopy "%PathToRepository%\buildtools\config\nginx" "%PathToAppFolder%\nginx\conf" /s /y /b /i
|
xcopy "%PathToRepository%\buildtools\config\nginx" "%PathToAppFolder%\nginx\conf" /s /y /b /i
|
||||||
xcopy "%PathToRepository%\buildtools\config\*" "%PathToAppFolder%\config" /y /b /i
|
xcopy "%PathToRepository%\buildtools\config\*" "%PathToAppFolder%\config" /y /b /i
|
||||||
|
@ -84,7 +84,11 @@ if ( $args.Count -ge 2 )
|
|||||||
|
|
||||||
$day = (Get-Date -Format "dddd").ToUpper().SubString(0, 3)
|
$day = (Get-Date -Format "dddd").ToUpper().SubString(0, 3)
|
||||||
$time = Get-Date -Format "HH:mm"
|
$time = Get-Date -Format "HH:mm"
|
||||||
cmd.exe /c "SCHTASKS /F /CREATE /SC WEEKLY /D $day /TN `"Certbot renew`" /TR `"${app}\letsencrypt\letsencrypt_cron.bat`" /ST $time"
|
$taskName = "Certbot renew"
|
||||||
|
$action = New-ScheduledTaskAction -Execute "cmd.exe" -Argument "/c `"${app}\letsencrypt\letsencrypt_cron.bat`""
|
||||||
|
$trigger = New-ScheduledTaskTrigger -Weekly -DaysOfWeek $day -At $time
|
||||||
|
|
||||||
|
Register-ScheduledTask -TaskName $taskName -Action $action -Trigger $trigger -Force
|
||||||
}
|
}
|
||||||
|
|
||||||
elseif ($args[0] -eq "-d" -or $args[0] -eq "--default") {
|
elseif ($args[0] -eq "-d" -or $args[0] -eq "--default") {
|
||||||
|
Loading…
Reference in New Issue
Block a user