mirror of
https://github.com/openimsdk/open-im-server.git
synced 2026-05-11 12:36:00 +08:00
Compare commits
35 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
| df03c76b85 | |||
| f185c981f7 | |||
| bf00aa74b9 | |||
| b49a8dd99d | |||
| ee04c156b5 | |||
| 646f1b9194 | |||
| 5041942563 | |||
| 649250b7c6 | |||
| 312c8ba9d6 | |||
| b36b69506f | |||
| 4be508a640 | |||
| 0207f1dab1 | |||
| bbac036d6c | |||
| 9baf1ffe82 | |||
| b5ef71f5c2 | |||
| 43919bc5fe | |||
| 0d03b46ac8 | |||
| a84f7bd217 | |||
| 9e8a389698 | |||
| a2110e416a | |||
| 0b612c13c6 | |||
| e7c7bf3bd1 | |||
| 9fb5ee5b55 | |||
| 3167f9943f | |||
| 598750e8c7 | |||
| 87f79d3cee | |||
| 7f44319b9b | |||
| c8dd598fbc | |||
| 678c23074f | |||
| 38a989b9fa | |||
| 758606f627 | |||
| 71f328ef94 | |||
| 9b94063d60 | |||
| 165eecf037 | |||
| 6890da44c9 |
@@ -1,5 +1,6 @@
|
|||||||
MONGO_IMAGE=mongo:7.0
|
MONGO_IMAGE=mongo:6.0.2
|
||||||
REDIS_IMAGE=redis:7.0.0
|
REDIS_IMAGE=redis:7.0.0
|
||||||
|
ZOOKEEPER_IMAGE=bitnami/zookeeper:3.8
|
||||||
KAFKA_IMAGE=bitnami/kafka:3.5.1
|
KAFKA_IMAGE=bitnami/kafka:3.5.1
|
||||||
MINIO_IMAGE=minio/minio:RELEASE.2024-01-11T07-46-16Z
|
MINIO_IMAGE=minio/minio:RELEASE.2024-01-11T07-46-16Z
|
||||||
ETCD_IMAGE=quay.io/coreos/etcd:v3.5.13
|
ETCD_IMAGE=quay.io/coreos/etcd:v3.5.13
|
||||||
@@ -15,3 +16,4 @@ OPENIM_ADMIN_FRONT_IMAGE=openim/openim-admin-front:release-v1.8.2
|
|||||||
#OPENIM_ADMIN_FRONT_IMAGE=registry.cn-hangzhou.aliyuncs.com/openimsdk/openim-admin-front:release-v1.8.2
|
#OPENIM_ADMIN_FRONT_IMAGE=registry.cn-hangzhou.aliyuncs.com/openimsdk/openim-admin-front:release-v1.8.2
|
||||||
|
|
||||||
DATA_DIR=./
|
DATA_DIR=./
|
||||||
|
|
||||||
|
|||||||
@@ -1,78 +0,0 @@
|
|||||||
name: Release Changelog
|
|
||||||
|
|
||||||
on:
|
|
||||||
release:
|
|
||||||
types: [released]
|
|
||||||
|
|
||||||
permissions:
|
|
||||||
contents: write
|
|
||||||
pull-requests: write
|
|
||||||
|
|
||||||
jobs:
|
|
||||||
update-changelog:
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
steps:
|
|
||||||
- name: Checkout code
|
|
||||||
uses: actions/checkout@v4
|
|
||||||
|
|
||||||
- name: Run Go Changelog Generator
|
|
||||||
run: |
|
|
||||||
# Run the Go changelog generator, passing the release tag if available
|
|
||||||
if [ "${{ github.event.release.tag_name }}" = "latest" ]; then
|
|
||||||
go run tools/changelog/changelog.go > "${{ github.event.release.tag_name }}-changelog.md"
|
|
||||||
else
|
|
||||||
go run tools/changelog/changelog.go "${{ github.event.release.tag_name }}" > "${{ github.event.release.tag_name }}-changelog.md"
|
|
||||||
fi
|
|
||||||
|
|
||||||
- name: Handle changelog files
|
|
||||||
run: |
|
|
||||||
# Ensure that the CHANGELOG directory exists
|
|
||||||
mkdir -p CHANGELOG
|
|
||||||
|
|
||||||
# Extract Major.Minor version by removing the 'v' prefix from the tag name
|
|
||||||
TAG_NAME=${{ github.event.release.tag_name }}
|
|
||||||
CHANGELOG_VERSION_NUMBER=$(echo "$TAG_NAME" | sed 's/^v//' | grep -oP '^\d+\.\d+')
|
|
||||||
|
|
||||||
# Define the new changelog file path
|
|
||||||
CHANGELOG_FILENAME="CHANGELOG-$CHANGELOG_VERSION_NUMBER.md"
|
|
||||||
CHANGELOG_PATH="CHANGELOG/$CHANGELOG_FILENAME"
|
|
||||||
|
|
||||||
# Check if the changelog file for the current release already exists
|
|
||||||
if [ -f "$CHANGELOG_PATH" ]; then
|
|
||||||
# If the file exists, append the new changelog to the existing one
|
|
||||||
cat "$CHANGELOG_PATH" >> "${TAG_NAME}-changelog.md"
|
|
||||||
# Overwrite the existing changelog with the updated content
|
|
||||||
mv "${TAG_NAME}-changelog.md" "$CHANGELOG_PATH"
|
|
||||||
else
|
|
||||||
# If the changelog file doesn't exist, rename the temp changelog file to the new changelog file
|
|
||||||
mv "${TAG_NAME}-changelog.md" "$CHANGELOG_PATH"
|
|
||||||
|
|
||||||
# Ensure that README.md exists
|
|
||||||
if [ ! -f "CHANGELOG/README.md" ]; then
|
|
||||||
echo -e "# CHANGELOGs\n\n" > CHANGELOG/README.md
|
|
||||||
fi
|
|
||||||
|
|
||||||
# Add the new changelog entry at the top of the README.md
|
|
||||||
if ! grep -q "\[$CHANGELOG_FILENAME\]" CHANGELOG/README.md; then
|
|
||||||
sed -i "3i- [$CHANGELOG_FILENAME](./$CHANGELOG_FILENAME)" CHANGELOG/README.md
|
|
||||||
# Remove the extra newline character added by sed
|
|
||||||
# sed -i '4d' CHANGELOG/README.md
|
|
||||||
fi
|
|
||||||
fi
|
|
||||||
|
|
||||||
- name: Clean up
|
|
||||||
run: |
|
|
||||||
# Remove any temporary files that were created during the process
|
|
||||||
rm -f "${{ github.event.release.tag_name }}-changelog.md"
|
|
||||||
|
|
||||||
- name: Create Pull Request
|
|
||||||
uses: peter-evans/create-pull-request@v7.0.5
|
|
||||||
with:
|
|
||||||
token: ${{ secrets.GITHUB_TOKEN }}
|
|
||||||
commit-message: "Update CHANGELOG for release ${{ github.event.release.tag_name }}"
|
|
||||||
title: "Update CHANGELOG for release ${{ github.event.release.tag_name }}"
|
|
||||||
body: "This PR updates the CHANGELOG files for release ${{ github.event.release.tag_name }}"
|
|
||||||
branch: changelog-${{ github.event.release.tag_name }}
|
|
||||||
base: main
|
|
||||||
delete-branch: true
|
|
||||||
labels: changelog
|
|
||||||
@@ -149,7 +149,7 @@ jobs:
|
|||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
strategy:
|
strategy:
|
||||||
matrix:
|
matrix:
|
||||||
go_version: ["1.22"]
|
go_version: ["1.21"]
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- name: Checkout Repository
|
- name: Checkout Repository
|
||||||
|
|||||||
@@ -1,84 +0,0 @@
|
|||||||
name: Update Version File on Release
|
|
||||||
|
|
||||||
on:
|
|
||||||
release:
|
|
||||||
types: [created]
|
|
||||||
|
|
||||||
jobs:
|
|
||||||
update-version:
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
env:
|
|
||||||
TAG_VERSION: ${{ github.event.release.tag_name }}
|
|
||||||
steps:
|
|
||||||
# Step 1: Checkout the original repository's code
|
|
||||||
- name: Checkout code
|
|
||||||
uses: actions/checkout@v4
|
|
||||||
with:
|
|
||||||
fetch-depth: 0
|
|
||||||
|
|
||||||
# Step 2: Set up Git with official account
|
|
||||||
- name: Set up Git
|
|
||||||
run: |
|
|
||||||
git config user.name "github-actions[bot]"
|
|
||||||
git config user.email "github-actions[bot]@users.noreply.github.com"
|
|
||||||
|
|
||||||
# Step 3: Check and delete existing tag
|
|
||||||
- name: Check and delete existing tag
|
|
||||||
run: |
|
|
||||||
if git rev-parse ${{ env.TAG_VERSION }} >/dev/null 2>&1; then
|
|
||||||
git tag -d ${{ env.TAG_VERSION }}
|
|
||||||
git push --delete origin ${{ env.TAG_VERSION }}
|
|
||||||
fi
|
|
||||||
|
|
||||||
# Step 4: Update version file
|
|
||||||
- name: Update version file
|
|
||||||
run: |
|
|
||||||
echo "${{ env.TAG_VERSION }}" > version/version
|
|
||||||
|
|
||||||
# Step 5: Commit and push changes
|
|
||||||
- name: Commit and push changes
|
|
||||||
env:
|
|
||||||
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
|
||||||
run: |
|
|
||||||
git add version/version
|
|
||||||
git commit -m "Update version to ${{ env.TAG_VERSION }}"
|
|
||||||
git push origin HEAD:${{ github.ref }}
|
|
||||||
|
|
||||||
# Step 6: Create and push tag
|
|
||||||
- name: Create and push tag
|
|
||||||
env:
|
|
||||||
GITHUB_TOKEN: ${{ secrets.GITHUB_TOKEN }}
|
|
||||||
run: |
|
|
||||||
git tag ${{ env.TAG_VERSION }}
|
|
||||||
git push origin ${{ env.TAG_VERSION }}
|
|
||||||
|
|
||||||
# Step 7: Find and Publish Draft Release
|
|
||||||
- name: Find and Publish Draft Release
|
|
||||||
uses: actions/github-script@v6
|
|
||||||
with:
|
|
||||||
github-token: ${{ secrets.GITHUB_TOKEN }}
|
|
||||||
script: |
|
|
||||||
// Get the list of releases
|
|
||||||
const releases = await github.rest.repos.listReleases({
|
|
||||||
owner: context.repo.owner,
|
|
||||||
repo: context.repo.repo
|
|
||||||
});
|
|
||||||
|
|
||||||
// Find the draft release where the title and tag_name are the same
|
|
||||||
const draftRelease = releases.data.find(release =>
|
|
||||||
release.draft && release.name === release.tag_name
|
|
||||||
);
|
|
||||||
|
|
||||||
if (draftRelease) {
|
|
||||||
// Publish the draft release using the release_id
|
|
||||||
await github.rest.repos.updateRelease({
|
|
||||||
owner: context.repo.owner,
|
|
||||||
repo: context.repo.repo,
|
|
||||||
release_id: draftRelease.id, // Use release_id
|
|
||||||
draft: false
|
|
||||||
});
|
|
||||||
|
|
||||||
core.info(`Draft Release ${draftRelease.tag_name} published successfully.`);
|
|
||||||
} else {
|
|
||||||
core.info("No matching draft release found.");
|
|
||||||
}
|
|
||||||
+3
-3
@@ -1,5 +1,5 @@
|
|||||||
# Use Go 1.22 Alpine as the base image for building the application
|
# Use Go 1.21 Alpine as the base image for building the application
|
||||||
FROM golang:1.22-alpine AS builder
|
FROM golang:1.21-alpine AS builder
|
||||||
|
|
||||||
# Define the base directory for the application as an environment variable
|
# Define the base directory for the application as an environment variable
|
||||||
ENV SERVER_DIR=/openim-server
|
ENV SERVER_DIR=/openim-server
|
||||||
@@ -22,7 +22,7 @@ RUN go install github.com/magefile/mage@v1.15.0
|
|||||||
RUN mage build
|
RUN mage build
|
||||||
|
|
||||||
# Using Alpine Linux with Go environment for the final image
|
# Using Alpine Linux with Go environment for the final image
|
||||||
FROM golang:1.22-alpine
|
FROM golang:1.21-alpine
|
||||||
|
|
||||||
# Install necessary packages, such as bash
|
# Install necessary packages, such as bash
|
||||||
RUN apk add --no-cache bash
|
RUN apk add --no-cache bash
|
||||||
|
|||||||
@@ -16,6 +16,7 @@
|
|||||||
[](https://www.bestpractices.dev/projects/8045)
|
[](https://www.bestpractices.dev/projects/8045)
|
||||||
[](https://github.com/openimsdk/open-im-server/issues?q=is%3Aissue+is%3Aopen+sort%3Aupdated-desc+label%3A%22good+first+issue%22)
|
[](https://github.com/openimsdk/open-im-server/issues?q=is%3Aissue+is%3Aopen+sort%3Aupdated-desc+label%3A%22good+first+issue%22)
|
||||||
[](https://golang.org/)
|
[](https://golang.org/)
|
||||||
|
[](https://gurubase.io/g/openim)
|
||||||
|
|
||||||
|
|
||||||
<p align="center">
|
<p align="center">
|
||||||
|
|||||||
@@ -8,8 +8,6 @@ database: openim_v3
|
|||||||
username: openIM
|
username: openIM
|
||||||
# Password for database authentication
|
# Password for database authentication
|
||||||
password: openIM123
|
password: openIM123
|
||||||
# Authentication source for database authentication, if use root user, set it to admin
|
|
||||||
authSource: openim_v3
|
|
||||||
# Maximum number of connections in the connection pool
|
# Maximum number of connections in the connection pool
|
||||||
maxPoolSize: 100
|
maxPoolSize: 100
|
||||||
# Maximum number of retry attempts for a failed database connection
|
# Maximum number of retry attempts for a failed database connection
|
||||||
|
|||||||
@@ -22,3 +22,6 @@ longConnSvr:
|
|||||||
websocketMaxMsgLen: 4096
|
websocketMaxMsgLen: 4096
|
||||||
# WebSocket connection handshake timeout in seconds
|
# WebSocket connection handshake timeout in seconds
|
||||||
websocketTimeout: 10
|
websocketTimeout: 10
|
||||||
|
|
||||||
|
# 1: For Android, iOS, Windows, Mac, and web platforms, only one instance can be online at a time
|
||||||
|
multiLoginPolicy: 1
|
||||||
|
|||||||
+16
-16
@@ -10,32 +10,32 @@ prometheus:
|
|||||||
# Enable or disable Prometheus monitoring
|
# Enable or disable Prometheus monitoring
|
||||||
enable: true
|
enable: true
|
||||||
# List of ports that Prometheus listens on; these must match the number of rpc.ports to ensure correct monitoring setup
|
# List of ports that Prometheus listens on; these must match the number of rpc.ports to ensure correct monitoring setup
|
||||||
ports: [ 12170, 12171, 12172, 12173, 12174, 12175, 12176, 12177, 12178, 12179, 12180, 12182, 12183, 12184, 12185, 12186 ]
|
ports: [ 12170, 12171, 12172, 12173, 12174, 12175, 12176, 12177, 12178, 12179, 12180, 12181, 12182, 12183, 12184, 12185 ]
|
||||||
|
|
||||||
maxConcurrentWorkers: 3
|
maxConcurrentWorkers: 3
|
||||||
#Use geTui for offline push notifications, or choose fcm or jpush; corresponding configuration settings must be specified.
|
#Use geTui for offline push notifications, or choose fcm or jpns; corresponding configuration settings must be specified.
|
||||||
enable: geTui
|
enable: geTui
|
||||||
geTui:
|
geTui:
|
||||||
pushUrl: https://restapi.getui.com/v2/$appId
|
pushUrl: https://restapi.getui.com/v2/$appId
|
||||||
masterSecret:
|
masterSecret:
|
||||||
appKey:
|
appKey:
|
||||||
intent:
|
intent:
|
||||||
channelID:
|
channelID:
|
||||||
channelName:
|
channelName:
|
||||||
fcm:
|
fcm:
|
||||||
# Prioritize using file paths. If the file path is empty, use URL
|
# Prioritize using file paths. If the file path is empty, use URL
|
||||||
filePath: # File path is concatenated with the parameters passed in through - c(`mage` default pass in `config/`) and filePath.
|
filePath: # File path is concatenated with the parameters passed in through - c(`mage` default pass in `config/`) and filePath.
|
||||||
authURL: # Must start with https or http.
|
authURL: # Must start with https or http.
|
||||||
jpush:
|
jpns:
|
||||||
appKey:
|
appKey:
|
||||||
masterSecret:
|
masterSecret:
|
||||||
pushURL:
|
pushURL:
|
||||||
pushIntent:
|
pushIntent:
|
||||||
|
|
||||||
# iOS system push sound and badge count
|
# iOS system push sound and badge count
|
||||||
iosPush:
|
iosPush:
|
||||||
pushSound: xxx
|
pushSound: xxx
|
||||||
badgeCount: true
|
badgeCount: true
|
||||||
production: false
|
production: false
|
||||||
|
|
||||||
fullUserCache: true
|
fullUserCache: true
|
||||||
|
|||||||
@@ -46,7 +46,7 @@ scrape_configs:
|
|||||||
- job_name: openimserver-openim-push
|
- job_name: openimserver-openim-push
|
||||||
static_configs:
|
static_configs:
|
||||||
- targets: [ internal_ip:12170, internal_ip:12171, internal_ip:12172, internal_ip:12173, internal_ip:12174, internal_ip:12175, internal_ip:12176, internal_ip:12177 ]
|
- targets: [ internal_ip:12170, internal_ip:12171, internal_ip:12172, internal_ip:12173, internal_ip:12174, internal_ip:12175, internal_ip:12176, internal_ip:12177 ]
|
||||||
# - targets: [ internal_ip:12170, internal_ip:12171, internal_ip:12172, internal_ip:12173, internal_ip:12174, internal_ip:12175, internal_ip:12176, internal_ip:12177, internal_ip:12178, internal_ip:12179, internal_ip:12180, internal_ip:12182, internal_ip:12183, internal_ip:12184, internal_ip:12185, internal_ip:12186 ]
|
# - targets: [ internal_ip:12170, internal_ip:12171, internal_ip:12172, internal_ip:12173, internal_ip:12174, internal_ip:12175, internal_ip:12176, internal_ip:12177, internal_ip:12178, internal_ip:12179, internal_ip:12180, internal_ip:12181, internal_ip:12182, internal_ip:12183, internal_ip:12184, internal_ip:12185 ]
|
||||||
labels:
|
labels:
|
||||||
namespace: default
|
namespace: default
|
||||||
- job_name: openimserver-openim-rpc-auth
|
- job_name: openimserver-openim-rpc-auth
|
||||||
|
|||||||
+12
-1
@@ -15,4 +15,15 @@ imAdminUserID: [ imAdmin ]
|
|||||||
# 1: For Android, iOS, Windows, Mac, and web platforms, only one instance can be online at a time
|
# 1: For Android, iOS, Windows, Mac, and web platforms, only one instance can be online at a time
|
||||||
multiLogin:
|
multiLogin:
|
||||||
policy: 1
|
policy: 1
|
||||||
maxNumOneEnd: 30
|
maxNumOneEnd: 30
|
||||||
|
customizeLoginNum:
|
||||||
|
ios: 1
|
||||||
|
android: 1
|
||||||
|
windows: 1
|
||||||
|
osx: 1
|
||||||
|
web: 1
|
||||||
|
miniWeb: 1
|
||||||
|
linux: 1
|
||||||
|
aPad: 1
|
||||||
|
iPad: 1
|
||||||
|
admin: 1
|
||||||
|
|||||||
@@ -240,11 +240,11 @@ push:
|
|||||||
channelName: ${GETUI_CHANNEL_NAME}
|
channelName: ${GETUI_CHANNEL_NAME}
|
||||||
fcm:
|
fcm:
|
||||||
serviceAccount: "${FCM_SERVICE_ACCOUNT}"
|
serviceAccount: "${FCM_SERVICE_ACCOUNT}"
|
||||||
jpush:
|
jpns:
|
||||||
appKey: ${JPUSH_APP_KEY}
|
appKey: ${JPNS_APP_KEY}
|
||||||
masterSecret: ${JPUSH_MASTER_SECRET}
|
masterSecret: ${JPNS_MASTER_SECRET}
|
||||||
pushUrl: ${JPUSH_PUSH_URL}
|
pushUrl: ${JPNS_PUSH_URL}
|
||||||
pushIntent: ${JPUSH_PUSH_INTENT}
|
pushIntent: ${JPNS_PUSH_INTENT}
|
||||||
|
|
||||||
# App manager configuration
|
# App manager configuration
|
||||||
#
|
#
|
||||||
|
|||||||
+8
-28
@@ -8,35 +8,12 @@ services:
|
|||||||
ports:
|
ports:
|
||||||
- "37017:27017"
|
- "37017:27017"
|
||||||
container_name: mongo
|
container_name: mongo
|
||||||
command: >
|
command: ["/bin/bash", "-c", "/docker-entrypoint-initdb.d/mongo-init.sh; docker-entrypoint.sh mongod --wiredTigerCacheSizeGB 1 --auth"]
|
||||||
bash -c '
|
|
||||||
docker-entrypoint.sh mongod --wiredTigerCacheSizeGB $$wiredTigerCacheSizeGB --auth &
|
|
||||||
until mongosh -u $$MONGO_INITDB_ROOT_USERNAME -p $$MONGO_INITDB_ROOT_PASSWORD --authenticationDatabase admin --eval "db.runCommand({ ping: 1 })" &>/dev/null; do
|
|
||||||
echo "Waiting for MongoDB to start..."
|
|
||||||
sleep 1
|
|
||||||
done &&
|
|
||||||
mongosh -u $$MONGO_INITDB_ROOT_USERNAME -p $$MONGO_INITDB_ROOT_PASSWORD --authenticationDatabase admin --eval "
|
|
||||||
db = db.getSiblingDB(\"$$MONGO_INITDB_DATABASE\");
|
|
||||||
if (!db.getUser(\"$$MONGO_OPENIM_USERNAME\")) {
|
|
||||||
db.createUser({
|
|
||||||
user: \"$$MONGO_OPENIM_USERNAME\",
|
|
||||||
pwd: \"$$MONGO_OPENIM_PASSWORD\",
|
|
||||||
roles: [{role: \"readWrite\", db: \"$$MONGO_INITDB_DATABASE\"}]
|
|
||||||
});
|
|
||||||
print(\"User created successfully: \");
|
|
||||||
print(\"Username: $$MONGO_OPENIM_USERNAME\");
|
|
||||||
print(\"Password: $$MONGO_OPENIM_PASSWORD\");
|
|
||||||
print(\"Database: $$MONGO_INITDB_DATABASE\");
|
|
||||||
} else {
|
|
||||||
print(\"User already exists in database: $$MONGO_INITDB_DATABASE, Username: $$MONGO_OPENIM_USERNAME\");
|
|
||||||
}
|
|
||||||
" &&
|
|
||||||
tail -f /dev/null
|
|
||||||
'
|
|
||||||
volumes:
|
volumes:
|
||||||
- "${DATA_DIR}/components/mongodb/data/db:/data/db"
|
- "${DATA_DIR}/components/mongodb/data/db:/data/db"
|
||||||
- "${DATA_DIR}/components/mongodb/data/logs:/data/logs"
|
- "${DATA_DIR}/components/mongodb/data/logs:/data/logs"
|
||||||
- "${DATA_DIR}/components/mongodb/data/conf:/etc/mongo"
|
- "${DATA_DIR}/components/mongodb/data/conf:/etc/mongo"
|
||||||
|
- "./scripts/mongo-init.sh:/docker-entrypoint-initdb.d/mongo-init.sh:ro"
|
||||||
environment:
|
environment:
|
||||||
- TZ=Asia/Shanghai
|
- TZ=Asia/Shanghai
|
||||||
- wiredTigerCacheSizeGB=1
|
- wiredTigerCacheSizeGB=1
|
||||||
@@ -94,7 +71,10 @@ services:
|
|||||||
ports:
|
ports:
|
||||||
- "19094:9094"
|
- "19094:9094"
|
||||||
volumes:
|
volumes:
|
||||||
|
- ./scripts/create-topic.sh:/opt/bitnami/kafka/create-topic.sh
|
||||||
- "${DATA_DIR}/components/kafka:/bitnami/kafka"
|
- "${DATA_DIR}/components/kafka:/bitnami/kafka"
|
||||||
|
command: >
|
||||||
|
bash -c "/opt/bitnami/scripts/kafka/run.sh & /opt/bitnami/kafka/create-topic.sh; wait"
|
||||||
environment:
|
environment:
|
||||||
#KAFKA_HEAP_OPTS: "-Xms128m -Xmx256m"
|
#KAFKA_HEAP_OPTS: "-Xms128m -Xmx256m"
|
||||||
TZ: Asia/Shanghai
|
TZ: Asia/Shanghai
|
||||||
@@ -105,11 +85,10 @@ services:
|
|||||||
KAFKA_CFG_ADVERTISED_LISTENERS: PLAINTEXT://kafka:9092,EXTERNAL://localhost:19094
|
KAFKA_CFG_ADVERTISED_LISTENERS: PLAINTEXT://kafka:9092,EXTERNAL://localhost:19094
|
||||||
KAFKA_CFG_LISTENER_SECURITY_PROTOCOL_MAP: CONTROLLER:PLAINTEXT,EXTERNAL:PLAINTEXT,PLAINTEXT:PLAINTEXT
|
KAFKA_CFG_LISTENER_SECURITY_PROTOCOL_MAP: CONTROLLER:PLAINTEXT,EXTERNAL:PLAINTEXT,PLAINTEXT:PLAINTEXT
|
||||||
KAFKA_CFG_CONTROLLER_LISTENER_NAMES: CONTROLLER
|
KAFKA_CFG_CONTROLLER_LISTENER_NAMES: CONTROLLER
|
||||||
KAFKA_NUM_PARTITIONS: 8
|
|
||||||
KAFKA_CFG_AUTO_CREATE_TOPICS_ENABLE: "true"
|
|
||||||
networks:
|
networks:
|
||||||
- openim
|
- openim
|
||||||
|
|
||||||
|
|
||||||
minio:
|
minio:
|
||||||
image: "${MINIO_IMAGE}"
|
image: "${MINIO_IMAGE}"
|
||||||
ports:
|
ports:
|
||||||
@@ -145,7 +124,7 @@ services:
|
|||||||
- "11002:80"
|
- "11002:80"
|
||||||
networks:
|
networks:
|
||||||
- openim
|
- openim
|
||||||
|
|
||||||
# prometheus:
|
# prometheus:
|
||||||
# image: ${PROMETHEUS_IMAGE}
|
# image: ${PROMETHEUS_IMAGE}
|
||||||
# container_name: prometheus
|
# container_name: prometheus
|
||||||
@@ -192,3 +171,4 @@ services:
|
|||||||
# - ${DATA_DIR:-./}/components/grafana:/var/lib/grafana
|
# - ${DATA_DIR:-./}/components/grafana:/var/lib/grafana
|
||||||
# networks:
|
# networks:
|
||||||
# - openim
|
# - openim
|
||||||
|
|
||||||
|
|||||||
@@ -474,10 +474,10 @@ This section involves setting up additional configuration variables for Websocke
|
|||||||
| GETUI_CHANNEL_ID | [User Defined] | GeTui Channel ID |
|
| GETUI_CHANNEL_ID | [User Defined] | GeTui Channel ID |
|
||||||
| GETUI_CHANNEL_NAME | [User Defined] | GeTui Channel Name |
|
| GETUI_CHANNEL_NAME | [User Defined] | GeTui Channel Name |
|
||||||
| FCM_SERVICE_ACCOUNT | "x.json" | FCM Service Account |
|
| FCM_SERVICE_ACCOUNT | "x.json" | FCM Service Account |
|
||||||
| JPUSH_APP_KEY | [User Defined] | JPUSH Application Key |
|
| JPNS_APP_KEY | [User Defined] | JPNS Application Key |
|
||||||
| JPUSH_MASTER_SECRET | [User Defined] | JPUSH Master Secret |
|
| JPNS_MASTER_SECRET | [User Defined] | JPNS Master Secret |
|
||||||
| JPUSH_PUSH_URL | [User Defined] | JPUSH Push Notification URL |
|
| JPNS_PUSH_URL | [User Defined] | JPNS Push Notification URL |
|
||||||
| JPUSH_PUSH_INTENT | [User Defined] | JPUSH Push Intent |
|
| JPNS_PUSH_INTENT | [User Defined] | JPNS Push Intent |
|
||||||
| IM_ADMIN_USERID | "imAdmin" | IM Administrator ID |
|
| IM_ADMIN_USERID | "imAdmin" | IM Administrator ID |
|
||||||
| IM_ADMIN_NAME | "imAdmin" | IM Administrator Nickname |
|
| IM_ADMIN_NAME | "imAdmin" | IM Administrator Nickname |
|
||||||
| MULTILOGIN_POLICY | "1" | Multi-login Policy |
|
| MULTILOGIN_POLICY | "1" | Multi-login Policy |
|
||||||
|
|||||||
@@ -1,8 +1,6 @@
|
|||||||
module github.com/openimsdk/open-im-server/v3
|
module github.com/openimsdk/open-im-server/v3
|
||||||
|
|
||||||
go 1.22.0
|
go 1.21.2
|
||||||
|
|
||||||
toolchain go1.23.2
|
|
||||||
|
|
||||||
require (
|
require (
|
||||||
firebase.google.com/go/v4 v4.14.1
|
firebase.google.com/go/v4 v4.14.1
|
||||||
@@ -10,12 +8,12 @@ require (
|
|||||||
github.com/gin-gonic/gin v1.9.1
|
github.com/gin-gonic/gin v1.9.1
|
||||||
github.com/go-playground/validator/v10 v10.20.0
|
github.com/go-playground/validator/v10 v10.20.0
|
||||||
github.com/gogo/protobuf v1.3.2 // indirect
|
github.com/gogo/protobuf v1.3.2 // indirect
|
||||||
github.com/golang-jwt/jwt/v4 v4.5.1
|
github.com/golang-jwt/jwt/v4 v4.5.0
|
||||||
github.com/gorilla/websocket v1.5.1
|
github.com/gorilla/websocket v1.5.1
|
||||||
github.com/grpc-ecosystem/go-grpc-prometheus v1.2.0
|
github.com/grpc-ecosystem/go-grpc-prometheus v1.2.0
|
||||||
github.com/mitchellh/mapstructure v1.5.0
|
github.com/mitchellh/mapstructure v1.5.0
|
||||||
github.com/openimsdk/protocol v0.0.72-alpha.54
|
github.com/openimsdk/protocol v0.0.72-alpha.51
|
||||||
github.com/openimsdk/tools v0.0.50-alpha.32
|
github.com/openimsdk/tools v0.0.50-alpha.16
|
||||||
github.com/pkg/errors v0.9.1 // indirect
|
github.com/pkg/errors v0.9.1 // indirect
|
||||||
github.com/prometheus/client_golang v1.18.0
|
github.com/prometheus/client_golang v1.18.0
|
||||||
github.com/stretchr/testify v1.9.0
|
github.com/stretchr/testify v1.9.0
|
||||||
@@ -94,7 +92,7 @@ require (
|
|||||||
github.com/fsnotify/fsnotify v1.7.0 // indirect
|
github.com/fsnotify/fsnotify v1.7.0 // indirect
|
||||||
github.com/gabriel-vasile/mimetype v1.4.3 // indirect
|
github.com/gabriel-vasile/mimetype v1.4.3 // indirect
|
||||||
github.com/gin-contrib/sse v0.1.0 // indirect
|
github.com/gin-contrib/sse v0.1.0 // indirect
|
||||||
github.com/go-logr/logr v1.4.2 // indirect
|
github.com/go-logr/logr v1.4.1 // indirect
|
||||||
github.com/go-logr/stdr v1.2.2 // indirect
|
github.com/go-logr/stdr v1.2.2 // indirect
|
||||||
github.com/go-ole/go-ole v1.2.6 // indirect
|
github.com/go-ole/go-ole v1.2.6 // indirect
|
||||||
github.com/go-playground/universal-translator v0.18.1 // indirect
|
github.com/go-playground/universal-translator v0.18.1 // indirect
|
||||||
|
|||||||
@@ -126,8 +126,8 @@ github.com/gin-contrib/sse v0.1.0/go.mod h1:RHrZQHXnP2xjPF+u1gW/2HnVO7nvIa9PG3Gm
|
|||||||
github.com/gin-gonic/gin v1.9.1 h1:4idEAncQnU5cB7BeOkPtxjfCSye0AAm1R0RVIqJ+Jmg=
|
github.com/gin-gonic/gin v1.9.1 h1:4idEAncQnU5cB7BeOkPtxjfCSye0AAm1R0RVIqJ+Jmg=
|
||||||
github.com/gin-gonic/gin v1.9.1/go.mod h1:hPrL7YrpYKXt5YId3A/Tnip5kqbEAP+KLuI3SUcPTeU=
|
github.com/gin-gonic/gin v1.9.1/go.mod h1:hPrL7YrpYKXt5YId3A/Tnip5kqbEAP+KLuI3SUcPTeU=
|
||||||
github.com/go-logr/logr v1.2.2/go.mod h1:jdQByPbusPIv2/zmleS9BjJVeZ6kBagPoEUsqbVz/1A=
|
github.com/go-logr/logr v1.2.2/go.mod h1:jdQByPbusPIv2/zmleS9BjJVeZ6kBagPoEUsqbVz/1A=
|
||||||
github.com/go-logr/logr v1.4.2 h1:6pFjapn8bFcIbiKo3XT4j/BhANplGihG6tvd+8rYgrY=
|
github.com/go-logr/logr v1.4.1 h1:pKouT5E8xu9zeFC39JXRDukb6JFQPXM5p5I91188VAQ=
|
||||||
github.com/go-logr/logr v1.4.2/go.mod h1:9T104GzyrTigFIr8wt5mBrctHMim0Nb2HLGrmQ40KvY=
|
github.com/go-logr/logr v1.4.1/go.mod h1:9T104GzyrTigFIr8wt5mBrctHMim0Nb2HLGrmQ40KvY=
|
||||||
github.com/go-logr/stdr v1.2.2 h1:hSWxHoqTgW2S2qGc0LTAI563KZ5YKYRhT3MFKZMbjag=
|
github.com/go-logr/stdr v1.2.2 h1:hSWxHoqTgW2S2qGc0LTAI563KZ5YKYRhT3MFKZMbjag=
|
||||||
github.com/go-logr/stdr v1.2.2/go.mod h1:mMo/vtBO5dYbehREoey6XUKy/eSumjCCveDpRre4VKE=
|
github.com/go-logr/stdr v1.2.2/go.mod h1:mMo/vtBO5dYbehREoey6XUKy/eSumjCCveDpRre4VKE=
|
||||||
github.com/go-ole/go-ole v1.2.6 h1:/Fpf6oFPoeFik9ty7siob0G6Ke8QvQEuVcuChpwXzpY=
|
github.com/go-ole/go-ole v1.2.6 h1:/Fpf6oFPoeFik9ty7siob0G6Ke8QvQEuVcuChpwXzpY=
|
||||||
@@ -158,8 +158,8 @@ github.com/godbus/dbus/v5 v5.0.4/go.mod h1:xhWf0FNVPg57R7Z0UbKHbJfkEywrmjJnf7w5x
|
|||||||
github.com/gogo/protobuf v1.3.2 h1:Ov1cvc58UF3b5XjBnZv7+opcTcQFZebYjWzi34vdm4Q=
|
github.com/gogo/protobuf v1.3.2 h1:Ov1cvc58UF3b5XjBnZv7+opcTcQFZebYjWzi34vdm4Q=
|
||||||
github.com/gogo/protobuf v1.3.2/go.mod h1:P1XiOD3dCwIKUDQYPy72D8LYyHL2YPYrpS2s69NZV8Q=
|
github.com/gogo/protobuf v1.3.2/go.mod h1:P1XiOD3dCwIKUDQYPy72D8LYyHL2YPYrpS2s69NZV8Q=
|
||||||
github.com/golang-jwt/jwt/v4 v4.4.2/go.mod h1:m21LjoU+eqJr34lmDMbreY2eSTRJ1cv77w39/MY0Ch0=
|
github.com/golang-jwt/jwt/v4 v4.4.2/go.mod h1:m21LjoU+eqJr34lmDMbreY2eSTRJ1cv77w39/MY0Ch0=
|
||||||
github.com/golang-jwt/jwt/v4 v4.5.1 h1:JdqV9zKUdtaa9gdPlywC3aeoEsR681PlKC+4F5gQgeo=
|
github.com/golang-jwt/jwt/v4 v4.5.0 h1:7cYmW1XlMY7h7ii7UhUyChSgS5wUJEnm9uZVTGqOWzg=
|
||||||
github.com/golang-jwt/jwt/v4 v4.5.1/go.mod h1:m21LjoU+eqJr34lmDMbreY2eSTRJ1cv77w39/MY0Ch0=
|
github.com/golang-jwt/jwt/v4 v4.5.0/go.mod h1:m21LjoU+eqJr34lmDMbreY2eSTRJ1cv77w39/MY0Ch0=
|
||||||
github.com/golang/glog v0.0.0-20160126235308-23def4e6c14b/go.mod h1:SBH7ygxi8pfUlaOkMMuAQtPIUF8ecWP5IEl/CR7VP2Q=
|
github.com/golang/glog v0.0.0-20160126235308-23def4e6c14b/go.mod h1:SBH7ygxi8pfUlaOkMMuAQtPIUF8ecWP5IEl/CR7VP2Q=
|
||||||
github.com/golang/groupcache v0.0.0-20200121045136-8c9f03a8e57e/go.mod h1:cIg4eruTrX1D+g88fzRXU5OdNfaM+9IcxsU14FzY7Hc=
|
github.com/golang/groupcache v0.0.0-20200121045136-8c9f03a8e57e/go.mod h1:cIg4eruTrX1D+g88fzRXU5OdNfaM+9IcxsU14FzY7Hc=
|
||||||
github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da h1:oI5xCqsCo564l8iNU+DwB5epxmsaqB+rhGL0m5jtYqE=
|
github.com/golang/groupcache v0.0.0-20210331224755-41bb18bfe9da h1:oI5xCqsCo564l8iNU+DwB5epxmsaqB+rhGL0m5jtYqE=
|
||||||
@@ -319,10 +319,10 @@ github.com/onsi/gomega v1.25.0 h1:Vw7br2PCDYijJHSfBOWhov+8cAnUf8MfMaIOV323l6Y=
|
|||||||
github.com/onsi/gomega v1.25.0/go.mod h1:r+zV744Re+DiYCIPRlYOTxn0YkOLcAnW8k1xXdMPGhM=
|
github.com/onsi/gomega v1.25.0/go.mod h1:r+zV744Re+DiYCIPRlYOTxn0YkOLcAnW8k1xXdMPGhM=
|
||||||
github.com/openimsdk/gomake v0.0.14-alpha.5 h1:VY9c5x515lTfmdhhPjMvR3BBRrRquAUCFsz7t7vbv7Y=
|
github.com/openimsdk/gomake v0.0.14-alpha.5 h1:VY9c5x515lTfmdhhPjMvR3BBRrRquAUCFsz7t7vbv7Y=
|
||||||
github.com/openimsdk/gomake v0.0.14-alpha.5/go.mod h1:PndCozNc2IsQIciyn9mvEblYWZwJmAI+06z94EY+csI=
|
github.com/openimsdk/gomake v0.0.14-alpha.5/go.mod h1:PndCozNc2IsQIciyn9mvEblYWZwJmAI+06z94EY+csI=
|
||||||
github.com/openimsdk/protocol v0.0.72-alpha.54 h1:opato7N4QjjRq/SHD54bDSVBpOEEDp1VLWVk5Os2A9s=
|
github.com/openimsdk/protocol v0.0.72-alpha.51 h1:G5Yjndp/FRyOJWhoQcSF2x2GvYiAIlqN0vjkvjUPycU=
|
||||||
github.com/openimsdk/protocol v0.0.72-alpha.54/go.mod h1:OZQA9FR55lseYoN2Ql1XAHYKHJGu7OMNkUbuekrKCM8=
|
github.com/openimsdk/protocol v0.0.72-alpha.51/go.mod h1:OZQA9FR55lseYoN2Ql1XAHYKHJGu7OMNkUbuekrKCM8=
|
||||||
github.com/openimsdk/tools v0.0.50-alpha.32 h1:JEsUFHFnaYg230TG+Ke3SUnaA2h44t4kABAzEdv5VZw=
|
github.com/openimsdk/tools v0.0.50-alpha.16 h1:bC1AQvJMuOHtZm8LZRvN8L5mH1Ws2VYdL+TLTs1iGSc=
|
||||||
github.com/openimsdk/tools v0.0.50-alpha.32/go.mod h1:r5U6RbxcR4xhKb2fhTmKGC9Yt5LcErHBVt3lhXQIHSo=
|
github.com/openimsdk/tools v0.0.50-alpha.16/go.mod h1:h1cYmfyaVtgFbKmb1Cfsl8XwUOMTt8ubVUQrdGtsUh4=
|
||||||
github.com/pelletier/go-toml/v2 v2.2.2 h1:aYUidT7k73Pcl9nb2gScu7NSrKCSHIDE89b3+6Wq+LM=
|
github.com/pelletier/go-toml/v2 v2.2.2 h1:aYUidT7k73Pcl9nb2gScu7NSrKCSHIDE89b3+6Wq+LM=
|
||||||
github.com/pelletier/go-toml/v2 v2.2.2/go.mod h1:1t835xjRzz80PqgE6HHgN2JOsmgYu/h4qDAS4n929Rs=
|
github.com/pelletier/go-toml/v2 v2.2.2/go.mod h1:1t835xjRzz80PqgE6HHgN2JOsmgYu/h4qDAS4n929Rs=
|
||||||
github.com/pierrec/lz4/v4 v4.1.21 h1:yOVMLb6qSIDP67pl/5F7RepeKYu/VmTyEXvuMI5d9mQ=
|
github.com/pierrec/lz4/v4 v4.1.21 h1:yOVMLb6qSIDP67pl/5F7RepeKYu/VmTyEXvuMI5d9mQ=
|
||||||
@@ -356,8 +356,8 @@ github.com/robfig/cron/v3 v3.0.1 h1:WdRxkvbJztn8LMz/QEvLN5sBU+xKpSqwwUO1Pjr4qDs=
|
|||||||
github.com/robfig/cron/v3 v3.0.1/go.mod h1:eQICP3HwyT7UooqI/z+Ov+PtYAWygg1TEWWzGIFLtro=
|
github.com/robfig/cron/v3 v3.0.1/go.mod h1:eQICP3HwyT7UooqI/z+Ov+PtYAWygg1TEWWzGIFLtro=
|
||||||
github.com/rogpeppe/go-internal v1.6.1/go.mod h1:xXDCJY+GAPziupqXw64V24skbSoqbTEfhy4qGm1nDQc=
|
github.com/rogpeppe/go-internal v1.6.1/go.mod h1:xXDCJY+GAPziupqXw64V24skbSoqbTEfhy4qGm1nDQc=
|
||||||
github.com/rogpeppe/go-internal v1.8.0/go.mod h1:WmiCO8CzOY8rg0OYDC4/i/2WRWAB6poM+XZ2dLUbcbE=
|
github.com/rogpeppe/go-internal v1.8.0/go.mod h1:WmiCO8CzOY8rg0OYDC4/i/2WRWAB6poM+XZ2dLUbcbE=
|
||||||
github.com/rogpeppe/go-internal v1.10.0 h1:TMyTOH3F/DB16zRVcYyreMH6GnZZrwQVAoYjRBZyWFQ=
|
github.com/rogpeppe/go-internal v1.11.0 h1:cWPaGQEPrBb5/AsnsZesgZZ9yb1OQ+GOISoDNXVBh4M=
|
||||||
github.com/rogpeppe/go-internal v1.10.0/go.mod h1:UQnix2H7Ngw/k4C5ijL5+65zddjncjaFoBhdsK/akog=
|
github.com/rogpeppe/go-internal v1.11.0/go.mod h1:ddIwULY96R17DhadqLgMfk9H9tvdUzkipdSkR5nkCZA=
|
||||||
github.com/rs/xid v1.5.0 h1:mKX4bl4iPYJtEIxp6CYiUuLQ/8DYMoz0PUdtGgMFRVc=
|
github.com/rs/xid v1.5.0 h1:mKX4bl4iPYJtEIxp6CYiUuLQ/8DYMoz0PUdtGgMFRVc=
|
||||||
github.com/rs/xid v1.5.0/go.mod h1:trrq9SKmegXys3aeAKXMUTdJsYXVwGY3RLcfgqegfbg=
|
github.com/rs/xid v1.5.0/go.mod h1:trrq9SKmegXys3aeAKXMUTdJsYXVwGY3RLcfgqegfbg=
|
||||||
github.com/russross/blackfriday/v2 v2.1.0/go.mod h1:+Rmxgy9KzJVeS9/2gXHxylqXiyQDYRxCVz55jmeOWTM=
|
github.com/russross/blackfriday/v2 v2.1.0/go.mod h1:+Rmxgy9KzJVeS9/2gXHxylqXiyQDYRxCVz55jmeOWTM=
|
||||||
|
|||||||
@@ -74,7 +74,7 @@ func newGinRouter(disCov discovery.SvcDiscoveryRegistry, config *Config) *gin.En
|
|||||||
case BestSpeed:
|
case BestSpeed:
|
||||||
r.Use(gzip.Gzip(gzip.BestSpeed))
|
r.Use(gzip.Gzip(gzip.BestSpeed))
|
||||||
}
|
}
|
||||||
r.Use(prommetricsGin(), gin.RecoveryWithWriter(gin.DefaultErrorWriter, mw.GinPanicErr), mw.CorsHandler(), mw.GinParseOperationID(), GinParseToken(authRpc))
|
r.Use(prommetricsGin(), gin.Recovery(), mw.CorsHandler(), mw.GinParseOperationID(), GinParseToken(authRpc))
|
||||||
u := NewUserApi(*userRpc)
|
u := NewUserApi(*userRpc)
|
||||||
m := NewMessageApi(messageRpc, userRpc, config.Share.IMAdminUserID)
|
m := NewMessageApi(messageRpc, userRpc, config.Share.IMAdminUserID)
|
||||||
j := jssdk.NewJSSdkApi(userRpc.Client, friendRpc.Client, groupRpc.Client, messageRpc.Client, conversationRpc.Client)
|
j := jssdk.NewJSSdkApi(userRpc.Client, friendRpc.Client, groupRpc.Client, messageRpc.Client, conversationRpc.Client)
|
||||||
|
|||||||
@@ -18,7 +18,6 @@ import (
|
|||||||
"context"
|
"context"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/openimsdk/tools/mw"
|
|
||||||
"runtime/debug"
|
"runtime/debug"
|
||||||
"sync"
|
"sync"
|
||||||
"sync/atomic"
|
"sync/atomic"
|
||||||
@@ -71,8 +70,6 @@ type Client struct {
|
|||||||
IsCompress bool `json:"isCompress"`
|
IsCompress bool `json:"isCompress"`
|
||||||
UserID string `json:"userID"`
|
UserID string `json:"userID"`
|
||||||
IsBackground bool `json:"isBackground"`
|
IsBackground bool `json:"isBackground"`
|
||||||
SDKType string `json:"sdkType"`
|
|
||||||
Encoder Encoder
|
|
||||||
ctx *UserConnContext
|
ctx *UserConnContext
|
||||||
longConnServer LongConnServer
|
longConnServer LongConnServer
|
||||||
closed atomic.Bool
|
closed atomic.Bool
|
||||||
@@ -98,17 +95,11 @@ func (c *Client) ResetClient(ctx *UserConnContext, conn LongConn, longConnServer
|
|||||||
c.closed.Store(false)
|
c.closed.Store(false)
|
||||||
c.closedErr = nil
|
c.closedErr = nil
|
||||||
c.token = ctx.GetToken()
|
c.token = ctx.GetToken()
|
||||||
c.SDKType = ctx.GetSDKType()
|
|
||||||
c.hbCtx, c.hbCancel = context.WithCancel(c.ctx)
|
c.hbCtx, c.hbCancel = context.WithCancel(c.ctx)
|
||||||
c.subLock = new(sync.Mutex)
|
c.subLock = new(sync.Mutex)
|
||||||
if c.subUserIDs != nil {
|
if c.subUserIDs != nil {
|
||||||
clear(c.subUserIDs)
|
clear(c.subUserIDs)
|
||||||
}
|
}
|
||||||
if c.SDKType == GoSDK {
|
|
||||||
c.Encoder = NewGobEncoder()
|
|
||||||
} else {
|
|
||||||
c.Encoder = NewJsonEncoder()
|
|
||||||
}
|
|
||||||
c.subUserIDs = make(map[string]struct{})
|
c.subUserIDs = make(map[string]struct{})
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -201,7 +192,7 @@ func (c *Client) handleMessage(message []byte) error {
|
|||||||
var binaryReq = getReq()
|
var binaryReq = getReq()
|
||||||
defer freeReq(binaryReq)
|
defer freeReq(binaryReq)
|
||||||
|
|
||||||
err := c.Encoder.Decode(message, binaryReq)
|
err := c.longConnServer.Decode(message, binaryReq)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@@ -348,7 +339,7 @@ func (c *Client) writeBinaryMsg(resp Resp) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
encodedBuf, err := c.Encoder.Encode(resp)
|
encodedBuf, err := c.longConnServer.Encode(resp)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@@ -376,11 +367,6 @@ func (c *Client) writeBinaryMsg(resp Resp) error {
|
|||||||
func (c *Client) activeHeartbeat(ctx context.Context) {
|
func (c *Client) activeHeartbeat(ctx context.Context) {
|
||||||
if c.PlatformID == constant.WebPlatformID {
|
if c.PlatformID == constant.WebPlatformID {
|
||||||
go func() {
|
go func() {
|
||||||
defer func() {
|
|
||||||
if r := recover(); r != nil {
|
|
||||||
mw.PanicStackToLog(ctx, r)
|
|
||||||
}
|
|
||||||
}()
|
|
||||||
log.ZDebug(ctx, "server initiative send heartbeat start.")
|
log.ZDebug(ctx, "server initiative send heartbeat start.")
|
||||||
ticker := time.NewTicker(pingPeriod)
|
ticker := time.NewTicker(pingPeriod)
|
||||||
defer ticker.Stop()
|
defer ticker.Stop()
|
||||||
@@ -452,8 +438,6 @@ func (c *Client) handlerTextMessage(b []byte) error {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
c.w.Lock()
|
|
||||||
defer c.w.Unlock()
|
|
||||||
if err := c.conn.SetWriteDeadline(writeWait); err != nil {
|
if err := c.conn.SetWriteDeadline(writeWait); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -27,12 +27,6 @@ const (
|
|||||||
GzipCompressionProtocol = "gzip"
|
GzipCompressionProtocol = "gzip"
|
||||||
BackgroundStatus = "isBackground"
|
BackgroundStatus = "isBackground"
|
||||||
SendResponse = "isMsgResp"
|
SendResponse = "isMsgResp"
|
||||||
SDKType = "sdkType"
|
|
||||||
)
|
|
||||||
|
|
||||||
const (
|
|
||||||
GoSDK = "go"
|
|
||||||
JsSDK = "js"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
|
|||||||
@@ -153,14 +153,6 @@ func (c *UserConnContext) GetCompression() bool {
|
|||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *UserConnContext) GetSDKType() string {
|
|
||||||
sdkType := c.Req.URL.Query().Get(SDKType)
|
|
||||||
if sdkType == "" {
|
|
||||||
sdkType = GoSDK
|
|
||||||
}
|
|
||||||
return sdkType
|
|
||||||
}
|
|
||||||
|
|
||||||
func (c *UserConnContext) ShouldSendResp() bool {
|
func (c *UserConnContext) ShouldSendResp() bool {
|
||||||
errResp, exists := c.Query(SendResponse)
|
errResp, exists := c.Query(SendResponse)
|
||||||
if exists {
|
if exists {
|
||||||
@@ -201,11 +193,7 @@ func (c *UserConnContext) ParseEssentialArgs() error {
|
|||||||
_, err := strconv.Atoi(platformIDStr)
|
_, err := strconv.Atoi(platformIDStr)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return servererrs.ErrConnArgsErr.WrapMsg("platformID is not int")
|
return servererrs.ErrConnArgsErr.WrapMsg("platformID is not int")
|
||||||
}
|
|
||||||
switch sdkType, _ := c.Query(SDKType); sdkType {
|
|
||||||
case "", GoSDK, JsSDK:
|
|
||||||
default:
|
|
||||||
return servererrs.ErrConnArgsErr.WrapMsg("sdkType is not go or js")
|
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -15,8 +15,6 @@
|
|||||||
package msggateway
|
package msggateway
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"bytes"
|
|
||||||
"encoding/gob"
|
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
|
|
||||||
"github.com/openimsdk/tools/errs"
|
"github.com/openimsdk/tools/errs"
|
||||||
@@ -29,46 +27,22 @@ type Encoder interface {
|
|||||||
|
|
||||||
type GobEncoder struct{}
|
type GobEncoder struct{}
|
||||||
|
|
||||||
func NewGobEncoder() Encoder {
|
func NewGobEncoder() *GobEncoder {
|
||||||
return GobEncoder{}
|
return &GobEncoder{}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (g GobEncoder) Encode(data any) ([]byte, error) {
|
func (g *GobEncoder) Encode(data any) ([]byte, error) {
|
||||||
var buff bytes.Buffer
|
|
||||||
enc := gob.NewEncoder(&buff)
|
|
||||||
if err := enc.Encode(data); err != nil {
|
|
||||||
return nil, errs.WrapMsg(err, "GobEncoder.Encode failed", "action", "encode")
|
|
||||||
}
|
|
||||||
return buff.Bytes(), nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (g GobEncoder) Decode(encodeData []byte, decodeData any) error {
|
|
||||||
buff := bytes.NewBuffer(encodeData)
|
|
||||||
dec := gob.NewDecoder(buff)
|
|
||||||
if err := dec.Decode(decodeData); err != nil {
|
|
||||||
return errs.WrapMsg(err, "GobEncoder.Decode failed", "action", "decode")
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
type JsonEncoder struct{}
|
|
||||||
|
|
||||||
func NewJsonEncoder() Encoder {
|
|
||||||
return JsonEncoder{}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (g JsonEncoder) Encode(data any) ([]byte, error) {
|
|
||||||
b, err := json.Marshal(data)
|
b, err := json.Marshal(data)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, errs.New("JsonEncoder.Encode failed", "action", "encode")
|
return nil, errs.New("Encoder.Encode failed", "action", "encode")
|
||||||
}
|
}
|
||||||
return b, nil
|
return b, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (g JsonEncoder) Decode(encodeData []byte, decodeData any) error {
|
func (g *GobEncoder) Decode(encodeData []byte, decodeData any) error {
|
||||||
err := json.Unmarshal(encodeData, decodeData)
|
err := json.Unmarshal(encodeData, decodeData)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return errs.New("JsonEncoder.Decode failed", "action", "decode")
|
return errs.New("Encoder.Decode failed", "action", "decode")
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -0,0 +1,48 @@
|
|||||||
|
package msggateway
|
||||||
|
|
||||||
|
import (
|
||||||
|
"testing"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestGobEncoder_Encode(t *testing.T) {
|
||||||
|
encoder := NewGobEncoder()
|
||||||
|
|
||||||
|
// 测试用例1: 编码 []byte 数据
|
||||||
|
inputData := []byte("test data")
|
||||||
|
encodedData, err := encoder.Encode(inputData)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("expected no error, got %v", err)
|
||||||
|
}
|
||||||
|
if string(encodedData) != string(inputData) {
|
||||||
|
t.Fatalf("expected encoded data to be '%s', got '%s'", inputData, encodedData)
|
||||||
|
}
|
||||||
|
|
||||||
|
// 测试用例2: 编码非 []byte 数据
|
||||||
|
nonByteData := "string data"
|
||||||
|
_, err = encoder.Encode(nonByteData)
|
||||||
|
if err == nil {
|
||||||
|
t.Fatalf("expected an error when encoding non-byte data, got none")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestGobEncoder_Decode(t *testing.T) {
|
||||||
|
encoder := NewGobEncoder()
|
||||||
|
|
||||||
|
// 测试用例1: 解码到 []byte 数据
|
||||||
|
encodedData := []byte("test data")
|
||||||
|
var decodedData []byte
|
||||||
|
err := encoder.Decode(encodedData, &decodedData)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("expected no error, got %v", err)
|
||||||
|
}
|
||||||
|
if string(decodedData) != string(encodedData) {
|
||||||
|
t.Fatalf("expected decoded data to be '%s', got '%s'", encodedData, decodedData)
|
||||||
|
}
|
||||||
|
|
||||||
|
// 测试用例2: 解码到非 []byte 数据
|
||||||
|
var nonByteData string
|
||||||
|
err = encoder.Decode(encodedData, &nonByteData)
|
||||||
|
if err == nil {
|
||||||
|
t.Fatalf("expected an error when decoding to non-byte data, got none")
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -83,11 +83,17 @@ func NewServer(rpcPort int, longConnServer LongConnServer, conf *Config, ready f
|
|||||||
return s
|
return s
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *Server) OnlinePushMsg(context context.Context, req *msggateway.OnlinePushMsgReq) (*msggateway.OnlinePushMsgResp, error) {
|
func (s *Server) OnlinePushMsg(
|
||||||
|
context context.Context,
|
||||||
|
req *msggateway.OnlinePushMsgReq,
|
||||||
|
) (*msggateway.OnlinePushMsgResp, error) {
|
||||||
panic("implement me")
|
panic("implement me")
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *Server) GetUsersOnlineStatus(ctx context.Context, req *msggateway.GetUsersOnlineStatusReq) (*msggateway.GetUsersOnlineStatusResp, error) {
|
func (s *Server) GetUsersOnlineStatus(
|
||||||
|
ctx context.Context,
|
||||||
|
req *msggateway.GetUsersOnlineStatusReq,
|
||||||
|
) (*msggateway.GetUsersOnlineStatusResp, error) {
|
||||||
if !authverify.IsAppManagerUid(ctx, s.config.Share.IMAdminUserID) {
|
if !authverify.IsAppManagerUid(ctx, s.config.Share.IMAdminUserID) {
|
||||||
return nil, errs.ErrNoPermission.WrapMsg("only app manager")
|
return nil, errs.ErrNoPermission.WrapMsg("only app manager")
|
||||||
}
|
}
|
||||||
@@ -215,7 +221,10 @@ func (s *Server) SuperGroupOnlineBatchPushOneMsg(ctx context.Context, req *msgga
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *Server) KickUserOffline(ctx context.Context, req *msggateway.KickUserOfflineReq) (*msggateway.KickUserOfflineResp, error) {
|
func (s *Server) KickUserOffline(
|
||||||
|
ctx context.Context,
|
||||||
|
req *msggateway.KickUserOfflineReq,
|
||||||
|
) (*msggateway.KickUserOfflineResp, error) {
|
||||||
for _, v := range req.KickUserIDList {
|
for _, v := range req.KickUserIDList {
|
||||||
clients, _, ok := s.LongConnServer.GetUserPlatformCons(v, int(req.PlatformID))
|
clients, _, ok := s.LongConnServer.GetUserPlatformCons(v, int(req.PlatformID))
|
||||||
if !ok {
|
if !ok {
|
||||||
|
|||||||
@@ -37,6 +37,7 @@ type LongConnServer interface {
|
|||||||
SetKickHandlerInfo(i *kickHandler)
|
SetKickHandlerInfo(i *kickHandler)
|
||||||
SubUserOnlineStatus(ctx context.Context, client *Client, data *Req) ([]byte, error)
|
SubUserOnlineStatus(ctx context.Context, client *Client, data *Req) ([]byte, error)
|
||||||
Compressor
|
Compressor
|
||||||
|
Encoder
|
||||||
MessageHandler
|
MessageHandler
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -60,7 +61,7 @@ type WsServer struct {
|
|||||||
authClient *rpcclient.Auth
|
authClient *rpcclient.Auth
|
||||||
disCov discovery.SvcDiscoveryRegistry
|
disCov discovery.SvcDiscoveryRegistry
|
||||||
Compressor
|
Compressor
|
||||||
//Encoder
|
Encoder
|
||||||
MessageHandler
|
MessageHandler
|
||||||
webhookClient *webhook.Client
|
webhookClient *webhook.Client
|
||||||
}
|
}
|
||||||
@@ -134,6 +135,7 @@ func NewWsServer(msgGatewayConfig *Config, opts ...Option) *WsServer {
|
|||||||
clients: newUserMap(),
|
clients: newUserMap(),
|
||||||
subscription: newSubscription(),
|
subscription: newSubscription(),
|
||||||
Compressor: NewGzipCompressor(),
|
Compressor: NewGzipCompressor(),
|
||||||
|
Encoder: NewGobEncoder(),
|
||||||
webhookClient: webhook.NewWebhookClient(msgGatewayConfig.WebhooksConfig.URL),
|
webhookClient: webhook.NewWebhookClient(msgGatewayConfig.WebhooksConfig.URL),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -276,7 +278,14 @@ func (ws *WsServer) registerClient(client *Client) {
|
|||||||
|
|
||||||
wg.Wait()
|
wg.Wait()
|
||||||
|
|
||||||
log.ZDebug(client.ctx, "user online", "online user Num", ws.onlineUserNum.Load(), "online user conn Num", ws.onlineUserConnNum.Load())
|
log.ZDebug(
|
||||||
|
client.ctx,
|
||||||
|
"user online",
|
||||||
|
"online user Num",
|
||||||
|
ws.onlineUserNum.Load(),
|
||||||
|
"online user conn Num",
|
||||||
|
ws.onlineUserConnNum.Load(),
|
||||||
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
func getRemoteAdders(client []*Client) string {
|
func getRemoteAdders(client []*Client) string {
|
||||||
@@ -318,6 +327,11 @@ func (ws *WsServer) multiTerminalLoginChecker(clientOK bool, oldClients []*Clien
|
|||||||
|
|
||||||
switch ws.msgGatewayConfig.Share.MultiLogin.Policy {
|
switch ws.msgGatewayConfig.Share.MultiLogin.Policy {
|
||||||
case constant.DefalutNotKick:
|
case constant.DefalutNotKick:
|
||||||
|
case constant.WebAndOther:
|
||||||
|
if constant.PlatformIDToClass(newClient.PlatformID) == constant.WebPlatformStr {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
fallthrough
|
||||||
case constant.PCAndOther:
|
case constant.PCAndOther:
|
||||||
if constant.PlatformIDToClass(newClient.PlatformID) == constant.TerminalPC {
|
if constant.PlatformIDToClass(newClient.PlatformID) == constant.TerminalPC {
|
||||||
return
|
return
|
||||||
@@ -342,7 +356,7 @@ func (ws *WsServer) multiTerminalLoginChecker(clientOK bool, oldClients []*Clien
|
|||||||
log.ZWarn(newClient.ctx, "InvalidateToken err", err, "userID", newClient.UserID,
|
log.ZWarn(newClient.ctx, "InvalidateToken err", err, "userID", newClient.UserID,
|
||||||
"platformID", newClient.PlatformID)
|
"platformID", newClient.PlatformID)
|
||||||
}
|
}
|
||||||
case constant.AllLoginButSameClassKick:
|
case constant.PcMobileAndWeb:
|
||||||
clients, ok := ws.clients.GetAll(newClient.UserID)
|
clients, ok := ws.clients.GetAll(newClient.UserID)
|
||||||
if !ok {
|
if !ok {
|
||||||
return
|
return
|
||||||
@@ -356,6 +370,21 @@ func (ws *WsServer) multiTerminalLoginChecker(clientOK bool, oldClients []*Clien
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
kickTokenFunc(kickClients)
|
kickTokenFunc(kickClients)
|
||||||
|
|
||||||
|
case constant.SingleTerminalLogin:
|
||||||
|
clients, ok := ws.clients.GetAll(newClient.UserID)
|
||||||
|
if !ok {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
var (
|
||||||
|
kickClients []*Client
|
||||||
|
)
|
||||||
|
for _, client := range clients {
|
||||||
|
kickClients = append(kickClients, client)
|
||||||
|
}
|
||||||
|
kickTokenFunc(kickClients)
|
||||||
|
case constant.Customize:
|
||||||
|
// todo
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -136,11 +136,6 @@ func (m *MsgTransfer) Start(index int, config *Config) error {
|
|||||||
|
|
||||||
if config.MsgTransfer.Prometheus.Enable {
|
if config.MsgTransfer.Prometheus.Enable {
|
||||||
go func() {
|
go func() {
|
||||||
defer func() {
|
|
||||||
if r := recover(); r != nil {
|
|
||||||
mw.PanicStackToLog(m.ctx, r)
|
|
||||||
}
|
|
||||||
}()
|
|
||||||
prometheusPort, err := datautil.GetElemByIndex(config.MsgTransfer.Prometheus.Ports, index)
|
prometheusPort, err := datautil.GetElemByIndex(config.MsgTransfer.Prometheus.Ports, index)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
netErr = err
|
netErr = err
|
||||||
|
|||||||
@@ -19,7 +19,6 @@ import (
|
|||||||
"encoding/json"
|
"encoding/json"
|
||||||
"errors"
|
"errors"
|
||||||
"github.com/openimsdk/open-im-server/v3/pkg/common/prommetrics"
|
"github.com/openimsdk/open-im-server/v3/pkg/common/prommetrics"
|
||||||
"github.com/openimsdk/tools/mw"
|
|
||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
"sync"
|
"sync"
|
||||||
@@ -347,12 +346,6 @@ func (och *OnlineHistoryRedisConsumerHandler) handleNotification(ctx context.Con
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
func (och *OnlineHistoryRedisConsumerHandler) HandleUserHasReadSeqMessages(ctx context.Context) {
|
func (och *OnlineHistoryRedisConsumerHandler) HandleUserHasReadSeqMessages(ctx context.Context) {
|
||||||
defer func() {
|
|
||||||
if r := recover(); r != nil {
|
|
||||||
mw.PanicStackToLog(ctx, r)
|
|
||||||
}
|
|
||||||
}()
|
|
||||||
|
|
||||||
defer och.wg.Done()
|
defer och.wg.Done()
|
||||||
|
|
||||||
for msg := range och.conversationUserHasReadChan {
|
for msg := range och.conversationUserHasReadChan {
|
||||||
|
|||||||
@@ -29,6 +29,5 @@ type Dummy struct {
|
|||||||
|
|
||||||
func (d *Dummy) Push(ctx context.Context, userIDs []string, title, content string, opts *options.Opts) error {
|
func (d *Dummy) Push(ctx context.Context, userIDs []string, title, content string, opts *options.Opts) error {
|
||||||
log.ZDebug(ctx, "dummy push")
|
log.ZDebug(ctx, "dummy push")
|
||||||
log.ZWarn(ctx, "Dummy push", nil, "ps", "The offline push is not configured. To configure it, please go to config/openim-push.yml.")
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -15,7 +15,6 @@
|
|||||||
package body
|
package body
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"github.com/openimsdk/open-im-server/v3/internal/push/offlinepush/options"
|
|
||||||
"github.com/openimsdk/open-im-server/v3/pkg/common/config"
|
"github.com/openimsdk/open-im-server/v3/pkg/common/config"
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -27,44 +26,38 @@ type Notification struct {
|
|||||||
|
|
||||||
type Android struct {
|
type Android struct {
|
||||||
Alert string `json:"alert,omitempty"`
|
Alert string `json:"alert,omitempty"`
|
||||||
Title string `json:"title,omitempty"`
|
|
||||||
Intent struct {
|
Intent struct {
|
||||||
URL string `json:"url,omitempty"`
|
URL string `json:"url,omitempty"`
|
||||||
} `json:"intent,omitempty"`
|
} `json:"intent,omitempty"`
|
||||||
Extras map[string]string `json:"extras,omitempty"`
|
Extras Extras `json:"extras"`
|
||||||
}
|
}
|
||||||
type Ios struct {
|
type Ios struct {
|
||||||
Alert IosAlert `json:"alert,omitempty"`
|
Alert string `json:"alert,omitempty"`
|
||||||
Sound string `json:"sound,omitempty"`
|
Sound string `json:"sound,omitempty"`
|
||||||
Badge string `json:"badge,omitempty"`
|
Badge string `json:"badge,omitempty"`
|
||||||
Extras map[string]string `json:"extras,omitempty"`
|
Extras Extras `json:"extras"`
|
||||||
MutableContent bool `json:"mutable-content"`
|
MutableContent bool `json:"mutable-content"`
|
||||||
}
|
}
|
||||||
|
|
||||||
type IosAlert struct {
|
type Extras struct {
|
||||||
Title string `json:"title,omitempty"`
|
ClientMsgID string `json:"clientMsgID"`
|
||||||
Body string `json:"body,omitempty"`
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (n *Notification) SetAlert(alert string, title string, opts *options.Opts) {
|
func (n *Notification) SetAlert(alert string) {
|
||||||
n.Alert = alert
|
n.Alert = alert
|
||||||
n.Android.Alert = alert
|
n.Android.Alert = alert
|
||||||
n.Android.Title = title
|
n.IOS.Alert = alert
|
||||||
n.IOS.Alert.Body = alert
|
n.IOS.Sound = "default"
|
||||||
n.IOS.Alert.Title = title
|
n.IOS.Badge = "+1"
|
||||||
n.IOS.Sound = opts.IOSPushSound
|
|
||||||
if opts.IOSBadgeCount {
|
|
||||||
n.IOS.Badge = "+1"
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (n *Notification) SetExtras(extras map[string]string) {
|
func (n *Notification) SetExtras(extras Extras) {
|
||||||
n.IOS.Extras = extras
|
n.IOS.Extras = extras
|
||||||
n.Android.Extras = extras
|
n.Android.Extras = extras
|
||||||
}
|
}
|
||||||
|
|
||||||
func (n *Notification) SetAndroidIntent(pushConf *config.Push) {
|
func (n *Notification) SetAndroidIntent(pushConf *config.Push) {
|
||||||
n.Android.Intent.URL = pushConf.JPush.PushIntent
|
n.Android.Intent.URL = pushConf.JPNS.PushIntent
|
||||||
}
|
}
|
||||||
|
|
||||||
func (n *Notification) IOSEnableMutableContent() {
|
func (n *Notification) IOSEnableMutableContent() {
|
||||||
|
|||||||
@@ -18,9 +18,9 @@ import (
|
|||||||
"context"
|
"context"
|
||||||
"encoding/base64"
|
"encoding/base64"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"github.com/openimsdk/open-im-server/v3/internal/push/offlinepush/options"
|
||||||
|
|
||||||
"github.com/openimsdk/open-im-server/v3/internal/push/offlinepush/jpush/body"
|
"github.com/openimsdk/open-im-server/v3/internal/push/offlinepush/jpush/body"
|
||||||
"github.com/openimsdk/open-im-server/v3/internal/push/offlinepush/options"
|
|
||||||
"github.com/openimsdk/open-im-server/v3/pkg/common/config"
|
"github.com/openimsdk/open-im-server/v3/pkg/common/config"
|
||||||
"github.com/openimsdk/tools/utils/httputil"
|
"github.com/openimsdk/tools/utils/httputil"
|
||||||
)
|
)
|
||||||
@@ -57,23 +57,17 @@ func (j *JPush) Push(ctx context.Context, userIDs []string, title, content strin
|
|||||||
var au body.Audience
|
var au body.Audience
|
||||||
au.SetAlias(userIDs)
|
au.SetAlias(userIDs)
|
||||||
var no body.Notification
|
var no body.Notification
|
||||||
extras := make(map[string]string)
|
var extras body.Extras
|
||||||
extras["ex"] = opts.Ex
|
|
||||||
if opts.Signal.ClientMsgID != "" {
|
if opts.Signal.ClientMsgID != "" {
|
||||||
extras["ClientMsgID"] = opts.Signal.ClientMsgID
|
extras.ClientMsgID = opts.Signal.ClientMsgID
|
||||||
}
|
}
|
||||||
no.IOSEnableMutableContent()
|
no.IOSEnableMutableContent()
|
||||||
no.SetExtras(extras)
|
no.SetExtras(extras)
|
||||||
no.SetAlert(content, title, opts)
|
no.SetAlert(title)
|
||||||
no.SetAndroidIntent(j.pushConf)
|
no.SetAndroidIntent(j.pushConf)
|
||||||
|
|
||||||
var msg body.Message
|
var msg body.Message
|
||||||
msg.SetMsgContent(content)
|
msg.SetMsgContent(content)
|
||||||
msg.SetTitle(title)
|
|
||||||
if opts.Signal.ClientMsgID != "" {
|
|
||||||
msg.SetExtras("ClientMsgID", opts.Signal.ClientMsgID)
|
|
||||||
}
|
|
||||||
msg.SetExtras("ex", opts.Ex)
|
|
||||||
var opt body.Options
|
var opt body.Options
|
||||||
opt.SetApnsProduction(j.pushConf.IOSPush.Production)
|
opt.SetApnsProduction(j.pushConf.IOSPush.Production)
|
||||||
var pushObj body.PushObj
|
var pushObj body.PushObj
|
||||||
@@ -82,26 +76,19 @@ func (j *JPush) Push(ctx context.Context, userIDs []string, title, content strin
|
|||||||
pushObj.SetNotification(&no)
|
pushObj.SetNotification(&no)
|
||||||
pushObj.SetMessage(&msg)
|
pushObj.SetMessage(&msg)
|
||||||
pushObj.SetOptions(&opt)
|
pushObj.SetOptions(&opt)
|
||||||
var resp map[string]any
|
var resp any
|
||||||
return j.request(ctx, pushObj, &resp, 5)
|
return j.request(ctx, pushObj, resp, 5)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (j *JPush) request(ctx context.Context, po body.PushObj, resp *map[string]any, timeout int) error {
|
func (j *JPush) request(ctx context.Context, po body.PushObj, resp any, timeout int) error {
|
||||||
err := j.httpClient.PostReturn(
|
return j.httpClient.PostReturn(
|
||||||
ctx,
|
ctx,
|
||||||
j.pushConf.JPush.PushURL,
|
j.pushConf.JPNS.PushURL,
|
||||||
map[string]string{
|
map[string]string{
|
||||||
"Authorization": j.getAuthorization(j.pushConf.JPush.AppKey, j.pushConf.JPush.MasterSecret),
|
"Authorization": j.getAuthorization(j.pushConf.JPNS.AppKey, j.pushConf.JPNS.MasterSecret),
|
||||||
},
|
},
|
||||||
po,
|
po,
|
||||||
resp,
|
resp,
|
||||||
timeout,
|
timeout,
|
||||||
)
|
)
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
if (*resp)["sendno"] != "0" {
|
|
||||||
return fmt.Errorf("jpush push failed %v", resp)
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -23,6 +23,8 @@ import (
|
|||||||
"github.com/openimsdk/open-im-server/v3/internal/push/offlinepush/options"
|
"github.com/openimsdk/open-im-server/v3/internal/push/offlinepush/options"
|
||||||
"github.com/openimsdk/open-im-server/v3/pkg/common/config"
|
"github.com/openimsdk/open-im-server/v3/pkg/common/config"
|
||||||
"github.com/openimsdk/open-im-server/v3/pkg/common/storage/cache"
|
"github.com/openimsdk/open-im-server/v3/pkg/common/storage/cache"
|
||||||
|
"github.com/openimsdk/tools/log"
|
||||||
|
"github.com/openimsdk/tools/mcontext"
|
||||||
"strings"
|
"strings"
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -49,6 +51,7 @@ func NewOfflinePusher(pushConf *config.Push, cache cache.ThirdCache, fcmConfigPa
|
|||||||
offlinePusher = jpush.NewClient(pushConf)
|
offlinePusher = jpush.NewClient(pushConf)
|
||||||
default:
|
default:
|
||||||
offlinePusher = dummy.NewClient()
|
offlinePusher = dummy.NewClient()
|
||||||
|
log.ZWarn(mcontext.WithMustInfoCtx([]string{"push start", "admin", "admin", ""}), "Unknown push config", nil)
|
||||||
}
|
}
|
||||||
return offlinePusher, nil
|
return offlinePusher, nil
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -73,7 +73,7 @@ func (o *OfflinePushConsumerHandler) getOfflinePushInfos(msg *sdkws.MsgData) (ti
|
|||||||
IsAtSelf bool `json:"isAtSelf"`
|
IsAtSelf bool `json:"isAtSelf"`
|
||||||
}
|
}
|
||||||
|
|
||||||
opts = &options.Opts{Signal: &options.Signal{ClientMsgID: msg.ClientMsgID}}
|
opts = &options.Opts{Signal: &options.Signal{}}
|
||||||
if msg.OfflinePushInfo != nil {
|
if msg.OfflinePushInfo != nil {
|
||||||
opts.IOSBadgeCount = msg.OfflinePushInfo.IOSBadgeCount
|
opts.IOSBadgeCount = msg.OfflinePushInfo.IOSBadgeCount
|
||||||
opts.IOSPushSound = msg.OfflinePushInfo.IOSPushSound
|
opts.IOSPushSound = msg.OfflinePushInfo.IOSPushSound
|
||||||
|
|||||||
@@ -4,10 +4,6 @@ import (
|
|||||||
"context"
|
"context"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
|
|
||||||
"math/rand"
|
|
||||||
"strconv"
|
|
||||||
"time"
|
|
||||||
|
|
||||||
"github.com/IBM/sarama"
|
"github.com/IBM/sarama"
|
||||||
"github.com/openimsdk/open-im-server/v3/internal/push/offlinepush"
|
"github.com/openimsdk/open-im-server/v3/internal/push/offlinepush"
|
||||||
"github.com/openimsdk/open-im-server/v3/internal/push/offlinepush/options"
|
"github.com/openimsdk/open-im-server/v3/internal/push/offlinepush/options"
|
||||||
@@ -31,6 +27,9 @@ import (
|
|||||||
"github.com/openimsdk/tools/utils/timeutil"
|
"github.com/openimsdk/tools/utils/timeutil"
|
||||||
"github.com/redis/go-redis/v9"
|
"github.com/redis/go-redis/v9"
|
||||||
"google.golang.org/protobuf/proto"
|
"google.golang.org/protobuf/proto"
|
||||||
|
"math/rand"
|
||||||
|
"strconv"
|
||||||
|
"time"
|
||||||
)
|
)
|
||||||
|
|
||||||
type ConsumerHandler struct {
|
type ConsumerHandler struct {
|
||||||
@@ -166,21 +165,17 @@ func (c *ConsumerHandler) Push2User(ctx context.Context, userIDs []string, msg *
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
needOfflinePushUserID := []string{msg.RecvID}
|
offlinePushUserID := []string{msg.RecvID}
|
||||||
var offlinePushUserID []string
|
|
||||||
|
|
||||||
//receiver offline push
|
//receiver offline push
|
||||||
if err = c.webhookBeforeOfflinePush(ctx, &c.config.WebhooksConfig.BeforeOfflinePush, needOfflinePushUserID, msg, &offlinePushUserID); err != nil {
|
if err = c.webhookBeforeOfflinePush(ctx, &c.config.WebhooksConfig.BeforeOfflinePush,
|
||||||
|
offlinePushUserID, msg, nil); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
log.ZInfo(ctx, "webhookBeforeOfflinePush end")
|
log.ZInfo(ctx, "webhookBeforeOfflinePush end")
|
||||||
|
err = c.offlinePushMsg(ctx, msg, offlinePushUserID)
|
||||||
if len(offlinePushUserID) > 0 {
|
|
||||||
needOfflinePushUserID = offlinePushUserID
|
|
||||||
}
|
|
||||||
err = c.offlinePushMsg(ctx, msg, needOfflinePushUserID)
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.ZWarn(ctx, "offlinePushMsg failed", err, "needOfflinePushUserID", needOfflinePushUserID, "msg", msg)
|
log.ZWarn(ctx, "offlinePushMsg failed", err, "offlinePushUserID", offlinePushUserID, "msg", msg)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -340,7 +335,6 @@ func (c *ConsumerHandler) groupMessagesHandler(ctx context.Context, groupID stri
|
|||||||
func (c *ConsumerHandler) offlinePushMsg(ctx context.Context, msg *sdkws.MsgData, offlinePushUserIDs []string) error {
|
func (c *ConsumerHandler) offlinePushMsg(ctx context.Context, msg *sdkws.MsgData, offlinePushUserIDs []string) error {
|
||||||
title, content, opts, err := c.getOfflinePushInfos(msg)
|
title, content, opts, err := c.getOfflinePushInfos(msg)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.ZError(ctx, "getOfflinePushInfos failed", err, "msg", msg)
|
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
err = c.offlinePusher.Push(ctx, offlinePushUserIDs, title, content, opts)
|
err = c.offlinePusher.Push(ctx, offlinePushUserIDs, title, content, opts)
|
||||||
@@ -370,7 +364,7 @@ func (c *ConsumerHandler) getOfflinePushInfos(msg *sdkws.MsgData) (title, conten
|
|||||||
IsAtSelf bool `json:"isAtSelf"`
|
IsAtSelf bool `json:"isAtSelf"`
|
||||||
}
|
}
|
||||||
|
|
||||||
opts = &options.Opts{Signal: &options.Signal{ClientMsgID: msg.ClientMsgID}}
|
opts = &options.Opts{Signal: &options.Signal{}}
|
||||||
if msg.OfflinePushInfo != nil {
|
if msg.OfflinePushInfo != nil {
|
||||||
opts.IOSBadgeCount = msg.OfflinePushInfo.IOSBadgeCount
|
opts.IOSBadgeCount = msg.OfflinePushInfo.IOSBadgeCount
|
||||||
opts.IOSPushSound = msg.OfflinePushInfo.IOSPushSound
|
opts.IOSPushSound = msg.OfflinePushInfo.IOSPushSound
|
||||||
|
|||||||
@@ -16,7 +16,6 @@ package auth
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"errors"
|
|
||||||
|
|
||||||
"github.com/openimsdk/open-im-server/v3/pkg/common/config"
|
"github.com/openimsdk/open-im-server/v3/pkg/common/config"
|
||||||
redis2 "github.com/openimsdk/open-im-server/v3/pkg/common/storage/cache/redis"
|
redis2 "github.com/openimsdk/open-im-server/v3/pkg/common/storage/cache/redis"
|
||||||
@@ -67,7 +66,6 @@ func Start(ctx context.Context, config *Config, client discovery.SvcDiscoveryReg
|
|||||||
config.Share.Secret,
|
config.Share.Secret,
|
||||||
config.RpcConfig.TokenPolicy.Expire,
|
config.RpcConfig.TokenPolicy.Expire,
|
||||||
config.Share.MultiLogin,
|
config.Share.MultiLogin,
|
||||||
config.Share.IMAdminUserID,
|
|
||||||
),
|
),
|
||||||
config: config,
|
config: config,
|
||||||
})
|
})
|
||||||
@@ -131,10 +129,6 @@ func (s *authServer) parseToken(ctx context.Context, tokensString string) (claim
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, errs.Wrap(err)
|
return nil, errs.Wrap(err)
|
||||||
}
|
}
|
||||||
isAdmin := authverify.IsManagerUserID(claims.UserID, s.config.Share.IMAdminUserID)
|
|
||||||
if isAdmin {
|
|
||||||
return claims, nil
|
|
||||||
}
|
|
||||||
m, err := s.authDatabase.GetTokensWithoutError(ctx, claims.UserID, claims.PlatformID)
|
m, err := s.authDatabase.GetTokensWithoutError(ctx, claims.UserID, claims.PlatformID)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
@@ -196,7 +190,7 @@ func (s *authServer) forceKickOff(ctx context.Context, userID string, platformID
|
|||||||
}
|
}
|
||||||
|
|
||||||
m, err := s.authDatabase.GetTokensWithoutError(ctx, userID, int(platformID))
|
m, err := s.authDatabase.GetTokensWithoutError(ctx, userID, int(platformID))
|
||||||
if err != nil && errors.Is(err, redis.Nil) {
|
if err != nil && err != redis.Nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
for k := range m {
|
for k := range m {
|
||||||
@@ -214,7 +208,7 @@ func (s *authServer) forceKickOff(ctx context.Context, userID string, platformID
|
|||||||
|
|
||||||
func (s *authServer) InvalidateToken(ctx context.Context, req *pbauth.InvalidateTokenReq) (*pbauth.InvalidateTokenResp, error) {
|
func (s *authServer) InvalidateToken(ctx context.Context, req *pbauth.InvalidateTokenReq) (*pbauth.InvalidateTokenResp, error) {
|
||||||
m, err := s.authDatabase.GetTokensWithoutError(ctx, req.UserID, int(req.PlatformID))
|
m, err := s.authDatabase.GetTokensWithoutError(ctx, req.UserID, int(req.PlatformID))
|
||||||
if err != nil && errors.Is(err, redis.Nil) {
|
if err != nil && err != redis.Nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
if m == nil {
|
if m == nil {
|
||||||
|
|||||||
@@ -261,35 +261,27 @@ func (c *conversationServer) SetConversations(ctx context.Context, req *pbconver
|
|||||||
|
|
||||||
setConversationFieldsFunc := func() {
|
setConversationFieldsFunc := func() {
|
||||||
if req.Conversation.RecvMsgOpt != nil {
|
if req.Conversation.RecvMsgOpt != nil {
|
||||||
conversation.RecvMsgOpt = req.Conversation.RecvMsgOpt.Value
|
|
||||||
m["recv_msg_opt"] = req.Conversation.RecvMsgOpt.Value
|
m["recv_msg_opt"] = req.Conversation.RecvMsgOpt.Value
|
||||||
}
|
}
|
||||||
if req.Conversation.AttachedInfo != nil {
|
if req.Conversation.AttachedInfo != nil {
|
||||||
conversation.AttachedInfo = req.Conversation.AttachedInfo.Value
|
|
||||||
m["attached_info"] = req.Conversation.AttachedInfo.Value
|
m["attached_info"] = req.Conversation.AttachedInfo.Value
|
||||||
}
|
}
|
||||||
if req.Conversation.Ex != nil {
|
if req.Conversation.Ex != nil {
|
||||||
conversation.Ex = req.Conversation.Ex.Value
|
|
||||||
m["ex"] = req.Conversation.Ex.Value
|
m["ex"] = req.Conversation.Ex.Value
|
||||||
}
|
}
|
||||||
if req.Conversation.IsPinned != nil {
|
if req.Conversation.IsPinned != nil {
|
||||||
conversation.IsPinned = req.Conversation.IsPinned.Value
|
|
||||||
m["is_pinned"] = req.Conversation.IsPinned.Value
|
m["is_pinned"] = req.Conversation.IsPinned.Value
|
||||||
}
|
}
|
||||||
if req.Conversation.GroupAtType != nil {
|
if req.Conversation.GroupAtType != nil {
|
||||||
conversation.GroupAtType = req.Conversation.GroupAtType.Value
|
|
||||||
m["group_at_type"] = req.Conversation.GroupAtType.Value
|
m["group_at_type"] = req.Conversation.GroupAtType.Value
|
||||||
}
|
}
|
||||||
if req.Conversation.MsgDestructTime != nil {
|
if req.Conversation.MsgDestructTime != nil {
|
||||||
conversation.MsgDestructTime = req.Conversation.MsgDestructTime.Value
|
|
||||||
m["msg_destruct_time"] = req.Conversation.MsgDestructTime.Value
|
m["msg_destruct_time"] = req.Conversation.MsgDestructTime.Value
|
||||||
}
|
}
|
||||||
if req.Conversation.IsMsgDestruct != nil {
|
if req.Conversation.IsMsgDestruct != nil {
|
||||||
conversation.IsMsgDestruct = req.Conversation.IsMsgDestruct.Value
|
|
||||||
m["is_msg_destruct"] = req.Conversation.IsMsgDestruct.Value
|
m["is_msg_destruct"] = req.Conversation.IsMsgDestruct.Value
|
||||||
}
|
}
|
||||||
if req.Conversation.BurnDuration != nil {
|
if req.Conversation.BurnDuration != nil {
|
||||||
conversation.BurnDuration = req.Conversation.BurnDuration.Value
|
|
||||||
m["burn_duration"] = req.Conversation.BurnDuration.Value
|
m["burn_duration"] = req.Conversation.BurnDuration.Value
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -16,7 +16,6 @@ package msg
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"errors"
|
|
||||||
|
|
||||||
cbapi "github.com/openimsdk/open-im-server/v3/pkg/callbackstruct"
|
cbapi "github.com/openimsdk/open-im-server/v3/pkg/callbackstruct"
|
||||||
"github.com/openimsdk/protocol/constant"
|
"github.com/openimsdk/protocol/constant"
|
||||||
@@ -109,7 +108,7 @@ func (m *msgServer) MarkMsgsAsRead(ctx context.Context, req *msg.MarkMsgsAsReadR
|
|||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
currentHasReadSeq, err := m.MsgDatabase.GetHasReadSeq(ctx, req.UserID, req.ConversationID)
|
currentHasReadSeq, err := m.MsgDatabase.GetHasReadSeq(ctx, req.UserID, req.ConversationID)
|
||||||
if err != nil && !errors.Is(err, redis.Nil) {
|
if err != nil && errs.Unwrap(err) != redis.Nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
if hasReadSeq > currentHasReadSeq {
|
if hasReadSeq > currentHasReadSeq {
|
||||||
@@ -137,7 +136,7 @@ func (m *msgServer) MarkConversationAsRead(ctx context.Context, req *msg.MarkCon
|
|||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
hasReadSeq, err := m.MsgDatabase.GetHasReadSeq(ctx, req.UserID, req.ConversationID)
|
hasReadSeq, err := m.MsgDatabase.GetHasReadSeq(ctx, req.UserID, req.ConversationID)
|
||||||
if err != nil && errors.Is(err, redis.Nil) {
|
if err != nil && errs.Unwrap(err) != redis.Nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
var seqs []int64
|
var seqs []int64
|
||||||
@@ -181,23 +180,14 @@ func (m *msgServer) MarkConversationAsRead(ctx context.Context, req *msg.MarkCon
|
|||||||
req.UserID, seqs, hasReadSeq)
|
req.UserID, seqs, hasReadSeq)
|
||||||
}
|
}
|
||||||
|
|
||||||
if conversation.ConversationType == constant.SingleChatType {
|
reqCall := &cbapi.CallbackGroupMsgReadReq{
|
||||||
reqCall := &cbapi.CallbackSingleMsgReadReq{
|
SendID: conversation.OwnerUserID,
|
||||||
ConversationID: conversation.ConversationID,
|
ReceiveID: req.UserID,
|
||||||
UserID: conversation.OwnerUserID,
|
UnreadMsgNum: req.HasReadSeq,
|
||||||
Seqs: req.Seqs,
|
ContentType: int64(conversation.ConversationType),
|
||||||
ContentType: conversation.ConversationType,
|
|
||||||
}
|
|
||||||
m.webhookAfterSingleMsgRead(ctx, &m.config.WebhooksConfig.AfterSingleMsgRead, reqCall)
|
|
||||||
} else if conversation.ConversationType == constant.ReadGroupChatType {
|
|
||||||
reqCall := &cbapi.CallbackGroupMsgReadReq{
|
|
||||||
SendID: conversation.OwnerUserID,
|
|
||||||
ReceiveID: req.UserID,
|
|
||||||
UnreadMsgNum: req.HasReadSeq,
|
|
||||||
ContentType: int64(conversation.ConversationType),
|
|
||||||
}
|
|
||||||
m.webhookAfterGroupMsgRead(ctx, &m.config.WebhooksConfig.AfterGroupMsgRead, reqCall)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
m.webhookAfterGroupMsgRead(ctx, &m.config.WebhooksConfig.AfterGroupMsgRead, reqCall)
|
||||||
return &msg.MarkConversationAsReadResp{}, nil
|
return &msg.MarkConversationAsReadResp{}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -16,7 +16,6 @@ package msg
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"github.com/openimsdk/tools/mw"
|
|
||||||
|
|
||||||
"github.com/openimsdk/open-im-server/v3/pkg/common/prommetrics"
|
"github.com/openimsdk/open-im-server/v3/pkg/common/prommetrics"
|
||||||
"github.com/openimsdk/open-im-server/v3/pkg/msgprocessor"
|
"github.com/openimsdk/open-im-server/v3/pkg/msgprocessor"
|
||||||
@@ -79,15 +78,8 @@ func (m *msgServer) sendMsgGroupChat(ctx context.Context, req *pbmsg.SendMsgReq)
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (m *msgServer) setConversationAtInfo(nctx context.Context, msg *sdkws.MsgData) {
|
func (m *msgServer) setConversationAtInfo(nctx context.Context, msg *sdkws.MsgData) {
|
||||||
|
|
||||||
log.ZDebug(nctx, "setConversationAtInfo", "msg", msg)
|
log.ZDebug(nctx, "setConversationAtInfo", "msg", msg)
|
||||||
|
|
||||||
defer func() {
|
|
||||||
if r := recover(); r != nil {
|
|
||||||
mw.PanicStackToLog(nctx, r)
|
|
||||||
}
|
|
||||||
}()
|
|
||||||
|
|
||||||
ctx := mcontext.NewCtx("@@@" + mcontext.GetOperationID(nctx))
|
ctx := mcontext.NewCtx("@@@" + mcontext.GetOperationID(nctx))
|
||||||
|
|
||||||
var atUserID []string
|
var atUserID []string
|
||||||
@@ -174,6 +166,9 @@ func (m *msgServer) sendMsgSingleChat(ctx context.Context, req *pbmsg.SendMsgReq
|
|||||||
prommetrics.SingleChatMsgProcessFailedCounter.Inc()
|
prommetrics.SingleChatMsgProcessFailedCounter.Inc()
|
||||||
return nil, nil
|
return nil, nil
|
||||||
} else {
|
} else {
|
||||||
|
if err = m.webhookBeforeSendSingleMsg(ctx, &m.config.WebhooksConfig.BeforeSendSingleMsg, req); err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
if err := m.webhookBeforeMsgModify(ctx, &m.config.WebhooksConfig.BeforeMsgModify, req); err != nil {
|
if err := m.webhookBeforeMsgModify(ctx, &m.config.WebhooksConfig.BeforeMsgModify, req); err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -16,15 +16,15 @@ package msg
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"errors"
|
|
||||||
pbmsg "github.com/openimsdk/protocol/msg"
|
pbmsg "github.com/openimsdk/protocol/msg"
|
||||||
|
"github.com/openimsdk/tools/errs"
|
||||||
"github.com/redis/go-redis/v9"
|
"github.com/redis/go-redis/v9"
|
||||||
"sort"
|
"sort"
|
||||||
)
|
)
|
||||||
|
|
||||||
func (m *msgServer) GetConversationMaxSeq(ctx context.Context, req *pbmsg.GetConversationMaxSeqReq) (*pbmsg.GetConversationMaxSeqResp, error) {
|
func (m *msgServer) GetConversationMaxSeq(ctx context.Context, req *pbmsg.GetConversationMaxSeqReq) (*pbmsg.GetConversationMaxSeqResp, error) {
|
||||||
maxSeq, err := m.MsgDatabase.GetMaxSeq(ctx, req.ConversationID)
|
maxSeq, err := m.MsgDatabase.GetMaxSeq(ctx, req.ConversationID)
|
||||||
if err != nil && !errors.Is(err, redis.Nil) {
|
if err != nil && errs.Unwrap(err) != redis.Nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
return &pbmsg.GetConversationMaxSeqResp{MaxSeq: maxSeq}, nil
|
return &pbmsg.GetConversationMaxSeqResp{MaxSeq: maxSeq}, nil
|
||||||
|
|||||||
@@ -139,11 +139,3 @@ func (m *msgServer) conversationAndGetRecvID(conversation *conversation.Conversa
|
|||||||
}
|
}
|
||||||
return ""
|
return ""
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *msgServer) AppendStreamMsg(ctx context.Context, req *msg.AppendStreamMsgReq) (*msg.AppendStreamMsgResp, error) {
|
|
||||||
return nil, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (m *msgServer) GetStreamMsg(ctx context.Context, req *msg.GetStreamMsgReq) (*msg.GetStreamMsgResp, error) {
|
|
||||||
return nil, nil
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -59,9 +59,6 @@ func (m *msgServer) messageVerification(ctx context.Context, data *msg.SendMsgRe
|
|||||||
data.MsgData.ContentType >= constant.NotificationBegin {
|
data.MsgData.ContentType >= constant.NotificationBegin {
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
if err := m.webhookBeforeSendSingleMsg(ctx, &m.config.WebhooksConfig.BeforeSendSingleMsg, data); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
black, err := m.FriendLocalCache.IsBlack(ctx, data.MsgData.SendID, data.MsgData.RecvID)
|
black, err := m.FriendLocalCache.IsBlack(ctx, data.MsgData.SendID, data.MsgData.RecvID)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
|
|||||||
@@ -69,7 +69,6 @@ type Mongo struct {
|
|||||||
Database string `mapstructure:"database"`
|
Database string `mapstructure:"database"`
|
||||||
Username string `mapstructure:"username"`
|
Username string `mapstructure:"username"`
|
||||||
Password string `mapstructure:"password"`
|
Password string `mapstructure:"password"`
|
||||||
AuthSource string `mapstructure:"authSource"`
|
|
||||||
MaxPoolSize int `mapstructure:"maxPoolSize"`
|
MaxPoolSize int `mapstructure:"maxPoolSize"`
|
||||||
MaxRetry int `mapstructure:"maxRetry"`
|
MaxRetry int `mapstructure:"maxRetry"`
|
||||||
}
|
}
|
||||||
@@ -213,12 +212,12 @@ type Push struct {
|
|||||||
FilePath string `mapstructure:"filePath"`
|
FilePath string `mapstructure:"filePath"`
|
||||||
AuthURL string `mapstructure:"authURL"`
|
AuthURL string `mapstructure:"authURL"`
|
||||||
} `mapstructure:"fcm"`
|
} `mapstructure:"fcm"`
|
||||||
JPush struct {
|
JPNS struct {
|
||||||
AppKey string `mapstructure:"appKey"`
|
AppKey string `mapstructure:"appKey"`
|
||||||
MasterSecret string `mapstructure:"masterSecret"`
|
MasterSecret string `mapstructure:"masterSecret"`
|
||||||
PushURL string `mapstructure:"pushURL"`
|
PushURL string `mapstructure:"pushURL"`
|
||||||
PushIntent string `mapstructure:"pushIntent"`
|
PushIntent string `mapstructure:"pushIntent"`
|
||||||
} `mapstructure:"jpush"`
|
} `mapstructure:"jpns"`
|
||||||
IOSPush struct {
|
IOSPush struct {
|
||||||
PushSound string `mapstructure:"pushSound"`
|
PushSound string `mapstructure:"pushSound"`
|
||||||
BadgeCount bool `mapstructure:"badgeCount"`
|
BadgeCount bool `mapstructure:"badgeCount"`
|
||||||
@@ -369,8 +368,20 @@ type Share struct {
|
|||||||
}
|
}
|
||||||
|
|
||||||
type MultiLogin struct {
|
type MultiLogin struct {
|
||||||
Policy int `mapstructure:"policy"`
|
Policy int `mapstructure:"policy"`
|
||||||
MaxNumOneEnd int `mapstructure:"maxNumOneEnd"`
|
MaxNumOneEnd int `mapstructure:"maxNumOneEnd"`
|
||||||
|
CustomizeLoginNum struct {
|
||||||
|
IOS int `mapstructure:"ios"`
|
||||||
|
Android int `mapstructure:"android"`
|
||||||
|
Windows int `mapstructure:"windows"`
|
||||||
|
OSX int `mapstructure:"osx"`
|
||||||
|
Web int `mapstructure:"web"`
|
||||||
|
MiniWeb int `mapstructure:"miniWeb"`
|
||||||
|
Linux int `mapstructure:"linux"`
|
||||||
|
APad int `mapstructure:"aPad"`
|
||||||
|
IPad int `mapstructure:"iPad"`
|
||||||
|
Admin int `mapstructure:"admin"`
|
||||||
|
} `mapstructure:"customizeLoginNum"`
|
||||||
}
|
}
|
||||||
|
|
||||||
type RpcRegisterName struct {
|
type RpcRegisterName struct {
|
||||||
@@ -479,7 +490,6 @@ func (m *Mongo) Build() *mongoutil.Config {
|
|||||||
Database: m.Database,
|
Database: m.Database,
|
||||||
Username: m.Username,
|
Username: m.Username,
|
||||||
Password: m.Password,
|
Password: m.Password,
|
||||||
AuthSource: m.AuthSource,
|
|
||||||
MaxPoolSize: m.MaxPoolSize,
|
MaxPoolSize: m.MaxPoolSize,
|
||||||
MaxRetry: m.MaxRetry,
|
MaxRetry: m.MaxRetry,
|
||||||
}
|
}
|
||||||
|
|||||||
+1
-1
@@ -38,7 +38,7 @@ const (
|
|||||||
func NewConversationRedis(rdb redis.UniversalClient, localCache *config.LocalCache, opts *rockscache.Options, db database.Conversation) cache.ConversationCache {
|
func NewConversationRedis(rdb redis.UniversalClient, localCache *config.LocalCache, opts *rockscache.Options, db database.Conversation) cache.ConversationCache {
|
||||||
batchHandler := NewBatchDeleterRedis(rdb, opts, []string{localCache.Conversation.Topic})
|
batchHandler := NewBatchDeleterRedis(rdb, opts, []string{localCache.Conversation.Topic})
|
||||||
c := localCache.Conversation
|
c := localCache.Conversation
|
||||||
log.ZDebug(context.Background(), "conversation local cache init", "Topic", c.Topic, "SlotNum", c.SlotNum, "SlotSize", c.SlotSize, "enable", c.Enable())
|
log.ZDebug(context.Background(), "black local cache init", "Topic", c.Topic, "SlotNum", c.SlotNum, "SlotSize", c.SlotSize, "enable", c.Enable())
|
||||||
return &ConversationRedisCache{
|
return &ConversationRedisCache{
|
||||||
BatchDeleter: batchHandler,
|
BatchDeleter: batchHandler,
|
||||||
rcClient: rockscache.NewClient(rdb, *opts),
|
rcClient: rockscache.NewClient(rdb, *opts),
|
||||||
|
|||||||
+14
@@ -1,3 +1,17 @@
|
|||||||
|
// Copyright © 2023 OpenIM. All rights reserved.
|
||||||
|
//
|
||||||
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
// you may not use this file except in compliance with the License.
|
||||||
|
// You may obtain a copy of the License at
|
||||||
|
//
|
||||||
|
// http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
//
|
||||||
|
// Unless required by applicable law or agreed to in writing, software
|
||||||
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
// See the License for the specific language governing permissions and
|
||||||
|
// limitations under the License.
|
||||||
|
|
||||||
package redis
|
package redis
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
|||||||
@@ -2,7 +2,6 @@ package controller
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
|
|
||||||
"github.com/golang-jwt/jwt/v4"
|
"github.com/golang-jwt/jwt/v4"
|
||||||
"github.com/openimsdk/open-im-server/v3/pkg/authverify"
|
"github.com/openimsdk/open-im-server/v3/pkg/authverify"
|
||||||
"github.com/openimsdk/open-im-server/v3/pkg/common/config"
|
"github.com/openimsdk/open-im-server/v3/pkg/common/config"
|
||||||
@@ -26,8 +25,9 @@ type AuthDatabase interface {
|
|||||||
}
|
}
|
||||||
|
|
||||||
type multiLoginConfig struct {
|
type multiLoginConfig struct {
|
||||||
Policy int
|
Policy int
|
||||||
MaxNumOneEnd int
|
MaxNumOneEnd int
|
||||||
|
CustomizeLoginNum map[int]int
|
||||||
}
|
}
|
||||||
|
|
||||||
type authDatabase struct {
|
type authDatabase struct {
|
||||||
@@ -35,15 +35,25 @@ type authDatabase struct {
|
|||||||
accessSecret string
|
accessSecret string
|
||||||
accessExpire int64
|
accessExpire int64
|
||||||
multiLogin multiLoginConfig
|
multiLogin multiLoginConfig
|
||||||
adminUserIDs []string
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func NewAuthDatabase(cache cache.TokenModel, accessSecret string, accessExpire int64, multiLogin config.MultiLogin, adminUserIDs []string) AuthDatabase {
|
func NewAuthDatabase(cache cache.TokenModel, accessSecret string, accessExpire int64, multiLogin config.MultiLogin) AuthDatabase {
|
||||||
return &authDatabase{cache: cache, accessSecret: accessSecret, accessExpire: accessExpire, multiLogin: multiLoginConfig{
|
return &authDatabase{cache: cache, accessSecret: accessSecret, accessExpire: accessExpire, multiLogin: multiLoginConfig{
|
||||||
Policy: multiLogin.Policy,
|
Policy: multiLogin.Policy,
|
||||||
MaxNumOneEnd: multiLogin.MaxNumOneEnd,
|
MaxNumOneEnd: multiLogin.MaxNumOneEnd,
|
||||||
}, adminUserIDs: adminUserIDs,
|
CustomizeLoginNum: map[int]int{
|
||||||
}
|
constant.IOSPlatformID: multiLogin.CustomizeLoginNum.IOS,
|
||||||
|
constant.AndroidPlatformID: multiLogin.CustomizeLoginNum.Android,
|
||||||
|
constant.WindowsPlatformID: multiLogin.CustomizeLoginNum.Windows,
|
||||||
|
constant.OSXPlatformID: multiLogin.CustomizeLoginNum.OSX,
|
||||||
|
constant.WebPlatformID: multiLogin.CustomizeLoginNum.Web,
|
||||||
|
constant.MiniWebPlatformID: multiLogin.CustomizeLoginNum.MiniWeb,
|
||||||
|
constant.LinuxPlatformID: multiLogin.CustomizeLoginNum.Linux,
|
||||||
|
constant.AndroidPadPlatformID: multiLogin.CustomizeLoginNum.APad,
|
||||||
|
constant.IPadPlatformID: multiLogin.CustomizeLoginNum.IPad,
|
||||||
|
constant.AdminPlatformID: multiLogin.CustomizeLoginNum.Admin,
|
||||||
|
},
|
||||||
|
}}
|
||||||
}
|
}
|
||||||
|
|
||||||
// If the result is empty.
|
// If the result is empty.
|
||||||
@@ -80,31 +90,27 @@ func (a *authDatabase) BatchSetTokenMapByUidPid(ctx context.Context, tokens []st
|
|||||||
|
|
||||||
// Create Token.
|
// Create Token.
|
||||||
func (a *authDatabase) CreateToken(ctx context.Context, userID string, platformID int) (string, error) {
|
func (a *authDatabase) CreateToken(ctx context.Context, userID string, platformID int) (string, error) {
|
||||||
isAdmin := authverify.IsManagerUserID(userID, a.adminUserIDs)
|
tokens, err := a.cache.GetAllTokensWithoutError(ctx, userID)
|
||||||
if !isAdmin {
|
if err != nil {
|
||||||
tokens, err := a.cache.GetAllTokensWithoutError(ctx, userID)
|
return "", err
|
||||||
|
}
|
||||||
|
deleteTokenKey, kickedTokenKey, err := a.checkToken(ctx, tokens, platformID)
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
if len(deleteTokenKey) != 0 {
|
||||||
|
err = a.cache.DeleteTokenByUidPid(ctx, userID, platformID, deleteTokenKey)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return "", err
|
return "", err
|
||||||
}
|
}
|
||||||
|
}
|
||||||
deleteTokenKey, kickedTokenKey, err := a.checkToken(ctx, tokens, platformID)
|
if len(kickedTokenKey) != 0 {
|
||||||
if err != nil {
|
for _, k := range kickedTokenKey {
|
||||||
return "", err
|
err := a.cache.SetTokenFlagEx(ctx, userID, platformID, k, constant.KickedToken)
|
||||||
}
|
|
||||||
if len(deleteTokenKey) != 0 {
|
|
||||||
err = a.cache.DeleteTokenByUidPid(ctx, userID, platformID, deleteTokenKey)
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return "", err
|
return "", err
|
||||||
}
|
}
|
||||||
}
|
log.ZDebug(ctx, "kicked token in create token", "token", k)
|
||||||
if len(kickedTokenKey) != 0 {
|
|
||||||
for _, k := range kickedTokenKey {
|
|
||||||
err := a.cache.SetTokenFlagEx(ctx, userID, platformID, k, constant.KickedToken)
|
|
||||||
if err != nil {
|
|
||||||
return "", err
|
|
||||||
}
|
|
||||||
log.ZDebug(ctx, "kicked token in create token", "token", k)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -115,12 +121,9 @@ func (a *authDatabase) CreateToken(ctx context.Context, userID string, platformI
|
|||||||
return "", errs.WrapMsg(err, "token.SignedString")
|
return "", errs.WrapMsg(err, "token.SignedString")
|
||||||
}
|
}
|
||||||
|
|
||||||
if !isAdmin {
|
if err = a.cache.SetTokenFlagEx(ctx, userID, platformID, tokenString, constant.NormalToken); err != nil {
|
||||||
if err = a.cache.SetTokenFlagEx(ctx, userID, platformID, tokenString, constant.NormalToken); err != nil {
|
return "", err
|
||||||
return "", err
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return tokenString, nil
|
return tokenString, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -168,8 +171,17 @@ func (a *authDatabase) checkToken(ctx context.Context, tokens map[int]map[string
|
|||||||
kickToken = append(kickToken, ts[len(ts)-1])
|
kickToken = append(kickToken, ts[len(ts)-1])
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
case constant.SingleTerminalLogin:
|
||||||
|
for _, ts := range loginTokenMap {
|
||||||
|
kickToken = append(kickToken, ts...)
|
||||||
|
}
|
||||||
|
case constant.WebAndOther:
|
||||||
|
unkickTerminal = constant.WebPlatformStr
|
||||||
|
fallthrough
|
||||||
case constant.PCAndOther:
|
case constant.PCAndOther:
|
||||||
unkickTerminal = constant.TerminalPC
|
if unkickTerminal == "" {
|
||||||
|
unkickTerminal = constant.TerminalPC
|
||||||
|
}
|
||||||
if constant.PlatformIDToClass(platformID) != unkickTerminal {
|
if constant.PlatformIDToClass(platformID) != unkickTerminal {
|
||||||
for plt, ts := range loginTokenMap {
|
for plt, ts := range loginTokenMap {
|
||||||
if constant.PlatformIDToClass(plt) != unkickTerminal {
|
if constant.PlatformIDToClass(plt) != unkickTerminal {
|
||||||
@@ -201,17 +213,17 @@ func (a *authDatabase) checkToken(ctx context.Context, tokens map[int]map[string
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
case constant.AllLoginButSameClassKick:
|
case constant.PcMobileAndWeb:
|
||||||
var (
|
var (
|
||||||
reserved = make(map[string]struct{})
|
reserved = make(map[string]bool)
|
||||||
)
|
)
|
||||||
|
|
||||||
for plt, ts := range loginTokenMap {
|
for plt, ts := range loginTokenMap {
|
||||||
if constant.PlatformIDToClass(plt) == constant.PlatformIDToClass(platformID) {
|
if constant.PlatformIDToClass(plt) == constant.PlatformIDToClass(platformID) {
|
||||||
kickToken = append(kickToken, ts...)
|
kickToken = append(kickToken, ts...)
|
||||||
} else {
|
} else {
|
||||||
if _, ok := reserved[constant.PlatformIDToClass(plt)]; !ok {
|
if !reserved[constant.PlatformIDToClass(plt)] {
|
||||||
reserved[constant.PlatformIDToClass(plt)] = struct{}{}
|
reserved[constant.PlatformIDToClass(plt)] = true
|
||||||
kickToken = append(kickToken, ts[:len(ts)-1]...)
|
kickToken = append(kickToken, ts[:len(ts)-1]...)
|
||||||
continue
|
continue
|
||||||
} else {
|
} else {
|
||||||
@@ -219,20 +231,36 @@ func (a *authDatabase) checkToken(ctx context.Context, tokens map[int]map[string
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
case constant.Customize:
|
||||||
|
if a.multiLogin.CustomizeLoginNum[platformID] <= 0 {
|
||||||
|
return nil, nil, errs.New("Do not allow login on this end").Wrap()
|
||||||
|
}
|
||||||
|
for plt, ts := range loginTokenMap {
|
||||||
|
l := len(ts)
|
||||||
|
if platformID == plt {
|
||||||
|
l++
|
||||||
|
}
|
||||||
|
// a.multiLogin.CustomizeLoginNum[platformID] must > 0
|
||||||
|
limit := min(a.multiLogin.CustomizeLoginNum[plt], a.multiLogin.MaxNumOneEnd)
|
||||||
|
if l > limit {
|
||||||
|
kickToken = append(kickToken, ts[:l-limit]...)
|
||||||
|
}
|
||||||
|
}
|
||||||
default:
|
default:
|
||||||
return nil, nil, errs.New("unknown multiLogin policy").Wrap()
|
return nil, nil, errs.New("unknown multiLogin policy").Wrap()
|
||||||
}
|
}
|
||||||
|
|
||||||
//var adminTokenMaxNum = a.multiLogin.MaxNumOneEnd
|
var adminTokenMaxNum = a.multiLogin.MaxNumOneEnd
|
||||||
//if a.multiLogin.Policy == constant.Customize {
|
if a.multiLogin.Policy == constant.Customize {
|
||||||
// adminTokenMaxNum = a.multiLogin.CustomizeLoginNum[constant.AdminPlatformID]
|
adminTokenMaxNum = a.multiLogin.CustomizeLoginNum[constant.AdminPlatformID]
|
||||||
//}
|
}
|
||||||
//l := len(adminToken)
|
l := len(adminToken)
|
||||||
//if platformID == constant.AdminPlatformID {
|
if platformID == constant.AdminPlatformID {
|
||||||
// l++
|
l++
|
||||||
//}
|
}
|
||||||
//if l > adminTokenMaxNum {
|
if l > adminTokenMaxNum {
|
||||||
// kickToken = append(kickToken, adminToken[:l-adminTokenMaxNum]...)
|
kickToken = append(kickToken, adminToken[:l-adminTokenMaxNum]...)
|
||||||
//}
|
}
|
||||||
return deleteToken, kickToken, nil
|
return deleteToken, kickToken, nil
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -372,7 +372,7 @@ func (db *commonMsgDatabase) getMsgBySeqsRange(ctx context.Context, userID strin
|
|||||||
// This ensures that their message retrieval starts from the point they joined.
|
// This ensures that their message retrieval starts from the point they joined.
|
||||||
func (db *commonMsgDatabase) GetMsgBySeqsRange(ctx context.Context, userID string, conversationID string, begin, end, num, userMaxSeq int64) (int64, int64, []*sdkws.MsgData, error) {
|
func (db *commonMsgDatabase) GetMsgBySeqsRange(ctx context.Context, userID string, conversationID string, begin, end, num, userMaxSeq int64) (int64, int64, []*sdkws.MsgData, error) {
|
||||||
userMinSeq, err := db.seqUser.GetUserMinSeq(ctx, conversationID, userID)
|
userMinSeq, err := db.seqUser.GetUserMinSeq(ctx, conversationID, userID)
|
||||||
if err != nil && !errors.Is(err, redis.Nil) {
|
if err != nil && errs.Unwrap(err) != redis.Nil {
|
||||||
return 0, 0, nil, err
|
return 0, 0, nil, err
|
||||||
}
|
}
|
||||||
minSeq, err := db.seqConversation.GetMinSeq(ctx, conversationID)
|
minSeq, err := db.seqConversation.GetMinSeq(ctx, conversationID)
|
||||||
@@ -443,11 +443,6 @@ func (db *commonMsgDatabase) GetMsgBySeqsRange(ctx context.Context, userID strin
|
|||||||
return 0, 0, nil, err
|
return 0, 0, nil, err
|
||||||
}
|
}
|
||||||
successMsgs = append(mongoMsgs, successMsgs...)
|
successMsgs = append(mongoMsgs, successMsgs...)
|
||||||
|
|
||||||
//_, err = db.msg.SetMessagesToCache(ctx, conversationID, mongoMsgs)
|
|
||||||
//if err != nil {
|
|
||||||
// return 0, 0, nil, err
|
|
||||||
//}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return minSeq, maxSeq, successMsgs, nil
|
return minSeq, maxSeq, successMsgs, nil
|
||||||
@@ -490,7 +485,7 @@ func (db *commonMsgDatabase) GetMsgBySeqs(ctx context.Context, userID string, co
|
|||||||
}
|
}
|
||||||
successMsgs, failedSeqs, err := db.msg.GetMessagesBySeq(ctx, conversationID, newSeqs)
|
successMsgs, failedSeqs, err := db.msg.GetMessagesBySeq(ctx, conversationID, newSeqs)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
if errors.Is(err, redis.Nil) {
|
if err != redis.Nil {
|
||||||
log.ZError(ctx, "get message from redis exception", err, "failedSeqs", failedSeqs, "conversationID", conversationID)
|
log.ZError(ctx, "get message from redis exception", err, "failedSeqs", failedSeqs, "conversationID", conversationID)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -505,11 +500,6 @@ func (db *commonMsgDatabase) GetMsgBySeqs(ctx context.Context, userID string, co
|
|||||||
}
|
}
|
||||||
|
|
||||||
successMsgs = append(successMsgs, mongoMsgs...)
|
successMsgs = append(successMsgs, mongoMsgs...)
|
||||||
|
|
||||||
//_, err = db.msg.SetMessagesToCache(ctx, conversationID, mongoMsgs)
|
|
||||||
//if err != nil {
|
|
||||||
// return 0, 0, nil, err
|
|
||||||
//}
|
|
||||||
}
|
}
|
||||||
return minSeq, maxSeq, successMsgs, nil
|
return minSeq, maxSeq, successMsgs, nil
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -76,7 +76,7 @@ func (g *GroupMgo) Take(ctx context.Context, groupID string) (group *model.Group
|
|||||||
|
|
||||||
func (g *GroupMgo) Search(ctx context.Context, keyword string, pagination pagination.Pagination) (total int64, groups []*model.Group, err error) {
|
func (g *GroupMgo) Search(ctx context.Context, keyword string, pagination pagination.Pagination) (total int64, groups []*model.Group, err error) {
|
||||||
// Define the sorting options
|
// Define the sorting options
|
||||||
opts := options.Find().SetSort(bson.D{{Key: "create_time", Value: -1}})
|
opts := options.Find().SetSort(bson.D{{Key: "created_at", Value: -1}})
|
||||||
|
|
||||||
// Perform the search with pagination and sorting
|
// Perform the search with pagination and sorting
|
||||||
return mongoutil.FindPage[*model.Group](ctx, g.coll, bson.M{
|
return mongoutil.FindPage[*model.Group](ctx, g.coll, bson.M{
|
||||||
|
|||||||
@@ -17,18 +17,12 @@ package rpccache
|
|||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"github.com/openimsdk/tools/mw"
|
|
||||||
|
|
||||||
"github.com/openimsdk/tools/log"
|
"github.com/openimsdk/tools/log"
|
||||||
"github.com/redis/go-redis/v9"
|
"github.com/redis/go-redis/v9"
|
||||||
)
|
)
|
||||||
|
|
||||||
func subscriberRedisDeleteCache(ctx context.Context, client redis.UniversalClient, channel string, del func(ctx context.Context, key ...string)) {
|
func subscriberRedisDeleteCache(ctx context.Context, client redis.UniversalClient, channel string, del func(ctx context.Context, key ...string)) {
|
||||||
defer func() {
|
|
||||||
if r := recover(); r != nil {
|
|
||||||
mw.PanicStackToLog(ctx, r)
|
|
||||||
}
|
|
||||||
}()
|
|
||||||
for message := range client.Subscribe(ctx, channel).Channel() {
|
for message := range client.Subscribe(ctx, channel).Channel() {
|
||||||
log.ZDebug(ctx, "subscriberRedisDeleteCache", "channel", channel, "payload", message.Payload)
|
log.ZDebug(ctx, "subscriberRedisDeleteCache", "channel", channel, "payload", message.Payload)
|
||||||
var keys []string
|
var keys []string
|
||||||
|
|||||||
Executable
+55
@@ -0,0 +1,55 @@
|
|||||||
|
#!/usr/bin/env bash
|
||||||
|
# Copyright © 2023 OpenIM. All rights reserved.
|
||||||
|
#
|
||||||
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
# you may not use this file except in compliance with the License.
|
||||||
|
# You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing, software
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
# See the License for the specific language governing permissions and
|
||||||
|
# limitations under the License.
|
||||||
|
|
||||||
|
# Wait for Kafka to be ready
|
||||||
|
|
||||||
|
KAFKA_SERVER=localhost:9092
|
||||||
|
|
||||||
|
MAX_ATTEMPTS=300
|
||||||
|
attempt_num=1
|
||||||
|
|
||||||
|
echo "Waiting for Kafka to be ready..."
|
||||||
|
|
||||||
|
until /opt/bitnami/kafka/bin/kafka-topics.sh --list --bootstrap-server $KAFKA_SERVER; do
|
||||||
|
echo "Attempt $attempt_num of $MAX_ATTEMPTS: Kafka not ready yet..."
|
||||||
|
if [ $attempt_num -eq $MAX_ATTEMPTS ]; then
|
||||||
|
echo "Kafka not ready after $MAX_ATTEMPTS attempts, exiting"
|
||||||
|
exit 1
|
||||||
|
fi
|
||||||
|
attempt_num=$((attempt_num+1))
|
||||||
|
sleep 1
|
||||||
|
done
|
||||||
|
|
||||||
|
echo "Kafka is ready. Creating topics..."
|
||||||
|
|
||||||
|
|
||||||
|
topics=("toRedis" "toMongo" "toPush" "toOfflinePush")
|
||||||
|
partitions=8
|
||||||
|
replicationFactor=1
|
||||||
|
|
||||||
|
for topic in "${topics[@]}"; do
|
||||||
|
if /opt/bitnami/kafka/bin/kafka-topics.sh --create \
|
||||||
|
--bootstrap-server $KAFKA_SERVER \
|
||||||
|
--replication-factor $replicationFactor \
|
||||||
|
--partitions $partitions \
|
||||||
|
--topic $topic
|
||||||
|
then
|
||||||
|
echo "Topic $topic created."
|
||||||
|
else
|
||||||
|
echo "Failed to create topic $topic."
|
||||||
|
fi
|
||||||
|
done
|
||||||
|
|
||||||
|
echo "All topics created."
|
||||||
Executable
+66
@@ -0,0 +1,66 @@
|
|||||||
|
# Copyright © 2023 OpenIM. All rights reserved.
|
||||||
|
#
|
||||||
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
# you may not use this file except in compliance with the License.
|
||||||
|
# You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing, software
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
# See the License for the specific language governing permissions and
|
||||||
|
# limitations under the License.
|
||||||
|
mongosh <<EOF
|
||||||
|
var maxRetries = 300;
|
||||||
|
var connected = false;
|
||||||
|
var rootUsername = '$MONGO_INITDB_ROOT_USERNAME';
|
||||||
|
var rootPassword = '$MONGO_INITDB_ROOT_PASSWORD';
|
||||||
|
var dbName = '$MONGO_INITDB_DATABASE';
|
||||||
|
var openimUsername = '$MONGO_OPENIM_USERNAME';
|
||||||
|
var openimPassword = '$MONGO_OPENIM_PASSWORD';
|
||||||
|
|
||||||
|
while (!connected && maxRetries > 0) {
|
||||||
|
try {
|
||||||
|
db = connect('mongodb://127.0.0.1:27017/admin');
|
||||||
|
var authResult = db.auth(rootUsername, rootPassword);
|
||||||
|
if (authResult) {
|
||||||
|
print('Authentication successful for root user: ' + rootUsername);
|
||||||
|
connected = true;
|
||||||
|
} else {
|
||||||
|
print('Authentication failed for root user: ' + rootUsername + ' with password: ' + rootPassword);
|
||||||
|
quit(1);
|
||||||
|
}
|
||||||
|
} catch (e) {
|
||||||
|
maxRetries--;
|
||||||
|
print('Connection failed, retrying... Remaining attempts: ' + maxRetries);
|
||||||
|
sleep(1000); // Sleep for 1 second
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (connected) {
|
||||||
|
db = db.getSiblingDB(dbName);
|
||||||
|
var createUserResult = db.createUser({
|
||||||
|
user: openimUsername,
|
||||||
|
pwd: openimPassword,
|
||||||
|
roles: [{
|
||||||
|
role: 'readWrite',
|
||||||
|
db: dbName
|
||||||
|
}]
|
||||||
|
});
|
||||||
|
|
||||||
|
if (createUserResult.ok == 1) {
|
||||||
|
print('User creation successful. User: ' + openimUsername + ', Database: ' + dbName);
|
||||||
|
} else {
|
||||||
|
print('User creation failed for user: ' + openimUsername + ' in database: ' + dbName);
|
||||||
|
quit(1);
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
print('Failed to connect to MongoDB after 300 retries.');
|
||||||
|
quit(1);
|
||||||
|
}
|
||||||
|
EOF
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
@@ -1,198 +0,0 @@
|
|||||||
package main
|
|
||||||
|
|
||||||
import (
|
|
||||||
"encoding/json"
|
|
||||||
"fmt"
|
|
||||||
"io"
|
|
||||||
"net/http"
|
|
||||||
"os"
|
|
||||||
"regexp"
|
|
||||||
"strings"
|
|
||||||
)
|
|
||||||
|
|
||||||
// You can specify a tag as a command line argument to generate the changelog for a specific version.
|
|
||||||
// Example: go run tools/changelog/changelog.go v0.0.33
|
|
||||||
// If no tag is provided, the latest release will be used.
|
|
||||||
|
|
||||||
// Setting repo owner and repo name by generate changelog
|
|
||||||
const (
|
|
||||||
repoOwner = "openimsdk"
|
|
||||||
repoName = "open-im-server"
|
|
||||||
)
|
|
||||||
|
|
||||||
// GitHubRepo struct represents the repo details.
|
|
||||||
type GitHubRepo struct {
|
|
||||||
Owner string
|
|
||||||
Repo string
|
|
||||||
FullChangelog string
|
|
||||||
}
|
|
||||||
|
|
||||||
// ReleaseData represents the JSON structure for release data.
|
|
||||||
type ReleaseData struct {
|
|
||||||
TagName string `json:"tag_name"`
|
|
||||||
Body string `json:"body"`
|
|
||||||
HtmlUrl string `json:"html_url"`
|
|
||||||
Published string `json:"published_at"`
|
|
||||||
}
|
|
||||||
|
|
||||||
// Method to classify and format release notes.
|
|
||||||
func (g *GitHubRepo) classifyReleaseNotes(body string) map[string][]string {
|
|
||||||
result := map[string][]string{
|
|
||||||
"feat": {},
|
|
||||||
"fix": {},
|
|
||||||
"chore": {},
|
|
||||||
"refactor": {},
|
|
||||||
"build": {},
|
|
||||||
"other": {},
|
|
||||||
}
|
|
||||||
|
|
||||||
// Regular expression to extract PR number and URL (case insensitive)
|
|
||||||
rePR := regexp.MustCompile(`(?i)in (https://github\.com/[^\s]+/pull/(\d+))`)
|
|
||||||
|
|
||||||
// Split the body into individual lines.
|
|
||||||
lines := strings.Split(body, "\n")
|
|
||||||
|
|
||||||
for _, line := range lines {
|
|
||||||
// Skip lines that contain "deps: Merge"
|
|
||||||
if strings.Contains(strings.ToLower(line), "deps: merge #") {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
|
|
||||||
// Use a regular expression to extract Full Changelog link and its title (case insensitive).
|
|
||||||
if strings.Contains(strings.ToLower(line), "**full changelog**") {
|
|
||||||
matches := regexp.MustCompile(`(?i)\*\*full changelog\*\*: (https://github\.com/[^\s]+/compare/([^\s]+))`).FindStringSubmatch(line)
|
|
||||||
if len(matches) > 2 {
|
|
||||||
// Format the Full Changelog link with title
|
|
||||||
g.FullChangelog = fmt.Sprintf("[%s](%s)", matches[2], matches[1])
|
|
||||||
}
|
|
||||||
continue // Skip further processing for this line.
|
|
||||||
}
|
|
||||||
|
|
||||||
if strings.HasPrefix(line, "*") {
|
|
||||||
var category string
|
|
||||||
|
|
||||||
// Use strings.ToLower to make the matching case insensitive
|
|
||||||
lowerLine := strings.ToLower(line)
|
|
||||||
|
|
||||||
// Determine the category based on the prefix (case insensitive).
|
|
||||||
if strings.HasPrefix(lowerLine, "* feat") {
|
|
||||||
category = "feat"
|
|
||||||
} else if strings.HasPrefix(lowerLine, "* fix") {
|
|
||||||
category = "fix"
|
|
||||||
} else if strings.HasPrefix(lowerLine, "* chore") {
|
|
||||||
category = "chore"
|
|
||||||
} else if strings.HasPrefix(lowerLine, "* refactor") {
|
|
||||||
category = "refactor"
|
|
||||||
} else if strings.HasPrefix(lowerLine, "* build") {
|
|
||||||
category = "build"
|
|
||||||
} else {
|
|
||||||
category = "other"
|
|
||||||
}
|
|
||||||
|
|
||||||
// Extract PR number and URL (case insensitive)
|
|
||||||
matches := rePR.FindStringSubmatch(line)
|
|
||||||
if len(matches) == 3 {
|
|
||||||
prURL := matches[1]
|
|
||||||
prNumber := matches[2]
|
|
||||||
// Format the line with the PR link and use original content for the final result
|
|
||||||
formattedLine := fmt.Sprintf("* %s [#%s](%s)", strings.Split(line, " by ")[0][2:], prNumber, prURL)
|
|
||||||
result[category] = append(result[category], formattedLine)
|
|
||||||
} else {
|
|
||||||
// If no PR link is found, just add the line as is
|
|
||||||
result[category] = append(result[category], line)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return result
|
|
||||||
}
|
|
||||||
|
|
||||||
// Method to generate the final changelog.
|
|
||||||
func (g *GitHubRepo) generateChangelog(tag, date, htmlURL, body string) string {
|
|
||||||
sections := g.classifyReleaseNotes(body)
|
|
||||||
|
|
||||||
// Convert ISO 8601 date to simpler format (YYYY-MM-DD)
|
|
||||||
formattedDate := date[:10]
|
|
||||||
|
|
||||||
// Changelog header with tag, date, and links.
|
|
||||||
changelog := fmt.Sprintf("## [%s](%s) \t(%s)\n\n", tag, htmlURL, formattedDate)
|
|
||||||
|
|
||||||
if len(sections["feat"]) > 0 {
|
|
||||||
changelog += "### New Features\n" + strings.Join(sections["feat"], "\n") + "\n\n"
|
|
||||||
}
|
|
||||||
if len(sections["fix"]) > 0 {
|
|
||||||
changelog += "### Bug Fixes\n" + strings.Join(sections["fix"], "\n") + "\n\n"
|
|
||||||
}
|
|
||||||
if len(sections["chore"]) > 0 {
|
|
||||||
changelog += "### Chores\n" + strings.Join(sections["chore"], "\n") + "\n\n"
|
|
||||||
}
|
|
||||||
if len(sections["refactor"]) > 0 {
|
|
||||||
changelog += "### Refactors\n" + strings.Join(sections["refactor"], "\n") + "\n\n"
|
|
||||||
}
|
|
||||||
if len(sections["build"]) > 0 {
|
|
||||||
changelog += "### Builds\n" + strings.Join(sections["build"], "\n") + "\n\n"
|
|
||||||
}
|
|
||||||
if len(sections["other"]) > 0 {
|
|
||||||
changelog += "### Others\n" + strings.Join(sections["other"], "\n") + "\n\n"
|
|
||||||
}
|
|
||||||
|
|
||||||
if g.FullChangelog != "" {
|
|
||||||
changelog += fmt.Sprintf("**Full Changelog**: %s\n", g.FullChangelog)
|
|
||||||
}
|
|
||||||
|
|
||||||
return changelog
|
|
||||||
}
|
|
||||||
|
|
||||||
// Method to fetch release data from GitHub API.
|
|
||||||
func (g *GitHubRepo) fetchReleaseData(version string) (*ReleaseData, error) {
|
|
||||||
var apiURL string
|
|
||||||
|
|
||||||
if version == "" {
|
|
||||||
// Fetch the latest release.
|
|
||||||
apiURL = fmt.Sprintf("https://api.github.com/repos/%s/%s/releases/latest", g.Owner, g.Repo)
|
|
||||||
} else {
|
|
||||||
// Fetch a specific version.
|
|
||||||
apiURL = fmt.Sprintf("https://api.github.com/repos/%s/%s/releases/tags/%s", g.Owner, g.Repo, version)
|
|
||||||
}
|
|
||||||
|
|
||||||
resp, err := http.Get(apiURL)
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
defer resp.Body.Close()
|
|
||||||
|
|
||||||
body, err := io.ReadAll(resp.Body)
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
|
|
||||||
var releaseData ReleaseData
|
|
||||||
err = json.Unmarshal(body, &releaseData)
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
|
|
||||||
return &releaseData, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func main() {
|
|
||||||
repo := &GitHubRepo{Owner: repoOwner, Repo: repoName}
|
|
||||||
|
|
||||||
// Get the version from command line arguments, if provided
|
|
||||||
var version string // Default is use latest
|
|
||||||
|
|
||||||
if len(os.Args) > 1 {
|
|
||||||
version = os.Args[1] // Use the provided version
|
|
||||||
}
|
|
||||||
|
|
||||||
// Fetch release data (either for latest or specific version)
|
|
||||||
releaseData, err := repo.fetchReleaseData(version)
|
|
||||||
if err != nil {
|
|
||||||
fmt.Println("Error fetching release data:", err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
// Generate and print the formatted changelog
|
|
||||||
changelog := repo.generateChangelog(releaseData.TagName, releaseData.Published, releaseData.HtmlUrl, releaseData.Body)
|
|
||||||
fmt.Println(changelog)
|
|
||||||
}
|
|
||||||
@@ -0,0 +1,308 @@
|
|||||||
|
// Copyright © 2023 OpenIM. All rights reserved.
|
||||||
|
//
|
||||||
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
// you may not use this file except in compliance with the License.
|
||||||
|
// You may obtain a copy of the License at
|
||||||
|
//
|
||||||
|
// http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
//
|
||||||
|
// Unless required by applicable law or agreed to in writing, software
|
||||||
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
// See the License for the specific language governing permissions and
|
||||||
|
// limitations under the License.
|
||||||
|
|
||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"log"
|
||||||
|
"os"
|
||||||
|
"os/exec"
|
||||||
|
"regexp"
|
||||||
|
"sort"
|
||||||
|
"strings"
|
||||||
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
mergeRequest = regexp.MustCompile(`Merge pull request #([\d]+)`)
|
||||||
|
webconsoleBump = regexp.MustCompile(regexp.QuoteMeta("bump(github.com/openshift/origin-web-console): ") + `([\w]+)`)
|
||||||
|
upstreamKube = regexp.MustCompile(`^UPSTREAM: (\d+)+:(.+)`)
|
||||||
|
upstreamRepo = regexp.MustCompile(`^UPSTREAM: ([\w/-]+): (\d+)+:(.+)`)
|
||||||
|
prefix = regexp.MustCompile(`^[\w-]: `)
|
||||||
|
|
||||||
|
assignments = []prefixAssignment{
|
||||||
|
{"cluster up", "cluster"},
|
||||||
|
{" pv ", "storage"},
|
||||||
|
{"haproxy", "router"},
|
||||||
|
{"router", "router"},
|
||||||
|
{"route", "route"},
|
||||||
|
{"authoriz", "auth"},
|
||||||
|
{"rbac", "auth"},
|
||||||
|
{"authent", "auth"},
|
||||||
|
{"reconcil", "auth"},
|
||||||
|
{"auth", "auth"},
|
||||||
|
{"role", "auth"},
|
||||||
|
{" dc ", "deploy"},
|
||||||
|
{"deployment", "deploy"},
|
||||||
|
{"rolling", "deploy"},
|
||||||
|
{"security context constr", "security"},
|
||||||
|
{"scc", "security"},
|
||||||
|
{"pipeline", "build"},
|
||||||
|
{"build", "build"},
|
||||||
|
{"registry", "registry"},
|
||||||
|
{"registries", "image"},
|
||||||
|
{"image", "image"},
|
||||||
|
{" arp ", "network"},
|
||||||
|
{" cni ", "network"},
|
||||||
|
{"egress", "network"},
|
||||||
|
{"network", "network"},
|
||||||
|
{"oc ", "cli"},
|
||||||
|
{"template", "template"},
|
||||||
|
{"etcd", "server"},
|
||||||
|
{"pod", "node"},
|
||||||
|
{"scripts/", "hack"},
|
||||||
|
{"e2e", "test"},
|
||||||
|
{"integration", "test"},
|
||||||
|
{"cluster", "cluster"},
|
||||||
|
{"master", "server"},
|
||||||
|
{"packages", "hack"},
|
||||||
|
{"api", "server"},
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
type prefixAssignment struct {
|
||||||
|
term string
|
||||||
|
prefix string
|
||||||
|
}
|
||||||
|
|
||||||
|
type commit struct {
|
||||||
|
short string
|
||||||
|
parents []string
|
||||||
|
message string
|
||||||
|
}
|
||||||
|
|
||||||
|
func contains(arr []string, value string) bool {
|
||||||
|
for _, s := range arr {
|
||||||
|
if s == value {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
func main() {
|
||||||
|
log.SetFlags(0)
|
||||||
|
if len(os.Args) != 3 {
|
||||||
|
log.Fatalf("Must specify two arguments, FROM and TO")
|
||||||
|
}
|
||||||
|
from := os.Args[1]
|
||||||
|
to := os.Args[2]
|
||||||
|
|
||||||
|
out, err := exec.Command("git", "log", "--topo-order", "--pretty=tformat:%h %p|%s", "--reverse", fmt.Sprintf("%s..%s", from, to)).CombinedOutput()
|
||||||
|
if err != nil {
|
||||||
|
log.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
hide := make(map[string]struct{})
|
||||||
|
var apiChanges []string
|
||||||
|
var webconsole []string
|
||||||
|
var commits []commit
|
||||||
|
var upstreams []commit
|
||||||
|
var bumps []commit
|
||||||
|
for _, line := range strings.Split(string(out), "\n") {
|
||||||
|
if len(strings.TrimSpace(line)) == 0 {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
parts := strings.SplitN(line, "|", 2)
|
||||||
|
hashes := strings.Split(parts[0], " ")
|
||||||
|
c := commit{short: hashes[0], parents: hashes[1:], message: parts[1]}
|
||||||
|
|
||||||
|
if strings.HasPrefix(c.message, "UPSTREAM: ") {
|
||||||
|
hide[c.short] = struct{}{}
|
||||||
|
upstreams = append(upstreams, c)
|
||||||
|
}
|
||||||
|
if strings.HasPrefix(c.message, "bump(") {
|
||||||
|
hide[c.short] = struct{}{}
|
||||||
|
bumps = append(bumps, c)
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(c.parents) == 1 {
|
||||||
|
commits = append(commits, c)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
matches := mergeRequest.FindStringSubmatch(line)
|
||||||
|
if len(matches) == 0 {
|
||||||
|
// this may have been a human pressing the merge button, we'll just record this as a direct push
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
// split the accumulated commits into any that are force merges (assumed to be the initial set due
|
||||||
|
// to --topo-order) from the PR commits as soon as we see any of our merge parents. Then print
|
||||||
|
// any of the force merges
|
||||||
|
var first int
|
||||||
|
for i := range commits {
|
||||||
|
first = i
|
||||||
|
if contains(c.parents, commits[i].short) {
|
||||||
|
first++
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
individual := commits[:first]
|
||||||
|
merged := commits[first:]
|
||||||
|
for _, commit := range individual {
|
||||||
|
if len(commit.parents) > 1 {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if _, ok := hide[commit.short]; ok {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
fmt.Printf("force-merge: %s %s\n", commit.message, commit.short)
|
||||||
|
}
|
||||||
|
|
||||||
|
// try to find either the PR title or the first commit title from the merge commit
|
||||||
|
out, err := exec.Command("git", "show", "--pretty=tformat:%b", c.short).CombinedOutput()
|
||||||
|
if err != nil {
|
||||||
|
log.Fatal(err)
|
||||||
|
}
|
||||||
|
var message string
|
||||||
|
para := strings.Split(string(out), "\n\n")
|
||||||
|
if len(para) > 0 && strings.HasPrefix(para[0], "Automatic merge from submit-queue") {
|
||||||
|
para = para[1:]
|
||||||
|
}
|
||||||
|
// this is no longer necessary with the submit queue in place
|
||||||
|
if len(para) > 0 && strings.HasPrefix(para[0], "Merged by ") {
|
||||||
|
para = para[1:]
|
||||||
|
}
|
||||||
|
// post submit-queue, the merge bot will add the PR title, which is usually pretty good
|
||||||
|
if len(para) > 0 {
|
||||||
|
message = strings.Split(para[0], "\n")[0]
|
||||||
|
}
|
||||||
|
if len(message) == 0 && len(merged) > 0 {
|
||||||
|
message = merged[0].message
|
||||||
|
}
|
||||||
|
if len(message) > 0 && len(merged) == 1 && message == merged[0].message {
|
||||||
|
merged = nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// try to calculate a prefix based on the diff
|
||||||
|
if len(message) > 0 && !prefix.MatchString(message) {
|
||||||
|
prefix, ok := findPrefixFor(message, merged)
|
||||||
|
if ok {
|
||||||
|
message = prefix + ": " + message
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// github merge
|
||||||
|
|
||||||
|
// has api changes
|
||||||
|
display := fmt.Sprintf("%s [\\#%s](https://github.com/openimsdk/Open-IM-Server/pull/%s)", message, matches[1], matches[1])
|
||||||
|
if hasFileChanges(c.short, "pkg/apistruct/") {
|
||||||
|
apiChanges = append(apiChanges, display)
|
||||||
|
}
|
||||||
|
|
||||||
|
var filtered []commit
|
||||||
|
for _, commit := range merged {
|
||||||
|
if _, ok := hide[commit.short]; ok {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
filtered = append(filtered, commit)
|
||||||
|
}
|
||||||
|
if len(filtered) > 0 {
|
||||||
|
fmt.Printf("- %s\n", display)
|
||||||
|
for _, commit := range filtered {
|
||||||
|
fmt.Printf(" - %s (%s)\n", commit.message, commit.short)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// stick the merge commit in at the beginning of the next list so we can anchor the previous parent
|
||||||
|
commits = []commit{c}
|
||||||
|
}
|
||||||
|
|
||||||
|
// chunk the bumps
|
||||||
|
var lines []string
|
||||||
|
for _, commit := range bumps {
|
||||||
|
if m := webconsoleBump.FindStringSubmatch(commit.message); len(m) > 0 {
|
||||||
|
webconsole = append(webconsole, m[1])
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
lines = append(lines, commit.message)
|
||||||
|
}
|
||||||
|
lines = sortAndUniq(lines)
|
||||||
|
for _, line := range lines {
|
||||||
|
fmt.Printf("- %s\n", line)
|
||||||
|
}
|
||||||
|
|
||||||
|
// chunk the upstreams
|
||||||
|
lines = nil
|
||||||
|
for _, commit := range upstreams {
|
||||||
|
lines = append(lines, commit.message)
|
||||||
|
}
|
||||||
|
lines = sortAndUniq(lines)
|
||||||
|
for _, line := range lines {
|
||||||
|
fmt.Printf("- %s\n", upstreamLinkify(line))
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(webconsole) > 0 {
|
||||||
|
fmt.Printf("- web: from %s^..%s\n", webconsole[0], webconsole[len(webconsole)-1])
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, apiChange := range apiChanges {
|
||||||
|
fmt.Printf(" - %s\n", apiChange)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func findPrefixFor(message string, commits []commit) (string, bool) {
|
||||||
|
message = strings.ToLower(message)
|
||||||
|
for _, m := range assignments {
|
||||||
|
if strings.Contains(message, m.term) {
|
||||||
|
return m.prefix, true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
for _, c := range commits {
|
||||||
|
if prefix, ok := findPrefixFor(c.message, nil); ok {
|
||||||
|
return prefix, ok
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return "", false
|
||||||
|
}
|
||||||
|
|
||||||
|
func hasFileChanges(commit string, prefixes ...string) bool {
|
||||||
|
out, err := exec.Command("git", "diff", "--name-only", fmt.Sprintf("%s^..%s", commit, commit)).CombinedOutput()
|
||||||
|
if err != nil {
|
||||||
|
log.Fatal(err)
|
||||||
|
}
|
||||||
|
for _, file := range strings.Split(string(out), "\n") {
|
||||||
|
for _, prefix := range prefixes {
|
||||||
|
if strings.HasPrefix(file, prefix) {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
|
func sortAndUniq(lines []string) []string {
|
||||||
|
sort.Strings(lines)
|
||||||
|
out := make([]string, 0, len(lines))
|
||||||
|
last := ""
|
||||||
|
for _, s := range lines {
|
||||||
|
if last == s {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
last = s
|
||||||
|
out = append(out, s)
|
||||||
|
}
|
||||||
|
return out
|
||||||
|
}
|
||||||
|
|
||||||
|
func upstreamLinkify(line string) string {
|
||||||
|
if m := upstreamKube.FindStringSubmatch(line); len(m) > 0 {
|
||||||
|
return fmt.Sprintf("UPSTREAM: [#%s](https://github.com/openimsdk/open-im-server/pull/%s):%s", m[1], m[1], m[2])
|
||||||
|
}
|
||||||
|
if m := upstreamRepo.FindStringSubmatch(line); len(m) > 0 {
|
||||||
|
return fmt.Sprintf("UPSTREAM: [%s#%s](https://github.com/%s/pull/%s):%s", m[1], m[2], m[1], m[2], m[3])
|
||||||
|
}
|
||||||
|
return line
|
||||||
|
}
|
||||||
@@ -66,7 +66,7 @@ func CheckMinIO(ctx context.Context, config *config.Minio) error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func CheckKafka(ctx context.Context, conf *config.Kafka) error {
|
func CheckKafka(ctx context.Context, conf *config.Kafka) error {
|
||||||
return kafka.CheckHealth(ctx, conf.Build())
|
return kafka.Check(ctx, conf.Build(), []string{conf.ToMongoTopic, conf.ToRedisTopic, conf.ToPushTopic, conf.ToOfflinePushTopic})
|
||||||
}
|
}
|
||||||
|
|
||||||
func initConfig(configDir string) (*config.Mongo, *config.Redis, *config.Kafka, *config.Minio, *config.Discovery, error) {
|
func initConfig(configDir string) (*config.Mongo, *config.Redis, *config.Kafka, *config.Minio, *config.Discovery, error) {
|
||||||
|
|||||||
+1
-1
@@ -1 +1 @@
|
|||||||
3.8.2
|
3.8.1
|
||||||
Reference in New Issue
Block a user