Daniel,
Thanks for an answer,
Just a question, what should be options for
debug=2
log_stderror=no
memdbg=5
memlog=5
mem_join=1
mem_safety=1
to get the most of
kamctl rpc corex.shm_summary ?
Cause when I'm setting
kamcmd corex.debug 5
I'm getting all of the list of fragments which is really
something not easy to read.
corex.shm_summary returns the following:
I see a lot of memory allocated to
core/tcp_main.c: tcpconn_new(1201) (around 100 MB count= 1225),
tls: tls_init.c: ser_malloc(364) (66 MB, count = 69815)
But still kamctl stats shmem showing 2GB of used memory.
Thanks in advance!
Hello,
first, the value for -M is too high, I cannot easily think of a case when one needs 1GB of private memory for each kamailio process.
You can try to run the rpc command shm.stats and look in the syslog for the report to check if it is different that what you get with mod.mem_stats.
Cheers,
Daniel
On 10.07.24 11:36, Ihor Olkhovskyi via sr-users wrote:
Hello!
I'm trying to get where all the SHM memory gone. For the moment I'm using these settings:-m 8192 -M 1024which means 8G of SHM meory.
Kamailio parameters are the following
fork=yes
children=8
tcp_children=12
enable_tls=yes
enable_sctp=no
tls_threads_mode=2
tcp_accept_no_cl=yes
tcp_max_connections=63536
tls_max_connections=63536
tcp_accept_aliases=no
tcp_async=yes
tcp_connect_timeout=10
tcp_conn_wq_max=63536
tcp_crlf_ping=yes
tcp_delayed_ack=yes
tcp_fd_cache=yes
tcp_keepalive=yes
tcp_keepcnt=3
tcp_keepidle=30
tcp_keepintvl=10
tcp_linger2=30
tcp_rd_buf_size=80000
tcp_send_timeout=10
tcp_wq_blk_size=2100
tcp_wq_max=10485760
open_files_limit=63536
And having ~1000 TLS/WSS clients I'm getting this after 3 days of running:{
"jsonrpc": "2.0",
"result": [
"shmem:fragments = 2796",
"shmem:free_size = 3737947072", (3.7 G)
"shmem:max_used_size = 4857418512", (4.8 G)
"shmem:real_used_size = 4851987520",
"shmem:total_size = 8589934592",
"shmem:used_size = 4838988096" (4.8 G)
],
"id": 984479
}Means half of SHM memory is gone
When running kamcmd mod.stats all shm I'm getting (output truncated for the largest values)
Module: core
{
...
tcpconn_new(1201): 130669280 (0.13 G)
...
Total: 131657632 (0.13 G)
}
Module: sl
{
...
Total: 7520
}
Module: siptrace
{
...
Total: 13520
}
Module: rr
{
Total: 0
}
Module: dialog
{
...
Total: 146080
}
Module: permissions
{
...
Total: 62144
}
Module: htable
{
....
Total: 3359552 (3.3 M)
}
Module: rtpengine
{
...
Total: 31456
}
Module: textopsx
{
Total: 0
}
Module: tsilo
{
...
Total: 75072
}
Module: tm
{
....
Total: 3459504 (3.4 M)
}
Module: usrloc
{
....
Total: 1217616
}
Module: pua_dialoginfo
{
...
Total: 8720
}
Module: pua
{
...
Total: 150848
}
Module: pike
{
...
Total: 9504
}
Module: websocket
{
wsconn_add(198): 54139856 (54 M)
...
Total: 54148096 (54 M)
}
Module: debugger
{
...
Total: 21872
}
Module: tmx
{
Total: 512
}
Module: kex
{
....
Total: 1888
}
Module: tls
{
...
ser_malloc(364): 87246112 (87 M)
...
Total: 87997168 (87 M)
}
Module: secfilter
{
...
Total: 768
}
Module: exec
{
...
Total: 16
}
Module: dispatcher
{
...
Total: 2992
}
Module: cfgutils
{
...
Total: 48
}
Module: app_python3
{
...
Total: 32
}
So, I'm guessing, where all the memory gone? Is there any way to get more detailed info on SHM?
Kamailio 5.8.2
--
Thanks in advance,Ihor
__________________________________________________________ Kamailio - Users Mailing List - Non Commercial Discussions To unsubscribe send an email to sr-users-leave@lists.kamailio.org Important: keep the mailing list in the recipients, do not reply only to the sender! Edit mailing list options or unsubscribe:-- Daniel-Constantin Mierla (@ asipto.com) twitter.com/miconda -- linkedin.com/in/miconda Kamailio Consultancy, Training and Development Services -- asipto.com