to get a personalized navigation.
to get a personalized navigation.
There are a lot of 5520-errors now. In addition we see "A connection was successfully established with the server, but then an error occurred during the pre-login handshake. (provider: SSL Provider, error: 0 - The wait operation timed out.)" sometimes.
The mitigation was finished yesterday (at least that is what it says on status.visma.com) ... but there is no doubt that something is not as it should be
Hi,
We have been able to identify the issue and we are currently applying a fix.
Please let us know if you are still seeing connection issues.
Thank you.
New strange error:
Url: /controller/api/v1/paymentMethod/
Content received:
{"message":"An error has occurred.","exceptionMessage":"The 'CompressedContent' type failed to serialize the response body for content type 'application/json; charset=utf-8'.","exceptionType":"System.InvalidOperationException","stackTrace":null,"innerException":{"message":"An error has occurred.","exceptionMessage":"Timeout performing EVAL (5000ms), next: EVAL, inst: 2, qu: 0, qs: 7, aw: False, rs: ReadAsync, ws: Idle, in: 0, serverEndpoint: 10.10.6.4:15005, mc: 1/1/0, mgr: 8 of 10 available, clientName: erp096701000036, PerfCounterHelperkeyHashSlot: 15367, IOCP: (Busy=0,Free=1000,Min=8,Max=1000), WORKER: (Busy=32,Free=32735,Min=8,Max=32767), v: 2.1.58.34321 (Please take a look at this article for some common client-side issues that can cause timeouts: https://stackexchange.github.io/StackExchange.Redis/Timeouts)","exceptionType":"StackExchange.Redis.RedisTimeoutException","stackTrace":" at StackExchange.Redis.ConnectionMultiplexer.ExecuteSyncImpl[T](Message message, ResultProcessor`1 processor, ServerEndPoint server) in /_/src/StackExchange.Redis/ConnectionMultiplexer.cs:line 2807\r\n at StackExchange.Redis.RedisBase.ExecuteSync[T](Message message, ResultProcessor`1 processor, ServerEndPoint server) in /_/src/StackExchange.Redis/RedisBase.cs:line 54\r\n at StackExchange.Redis.RedisDatabase.ScriptEvaluate(String script, RedisKey[] keys, RedisValue[] values, CommandFlags flags) in /_/src/StackExchange.Redis/RedisDatabase.cs:line 1153\r\n at Microsoft.Web.Redis.StackExchangeClientConnection.<>c__DisplayClass11_0.b__0()\r\n at Microsoft.Web.Redis.StackExchangeClientConnection.RetryForScriptNotFound(Func`1 redisOperation)\r\n at Microsoft.Web.Redis.StackExchangeClientConnection.RetryLogic(Func`1 redisOperation)\r\n at Microsoft.Web.Redis.StackExchangeClientConnection.Eval(String script, String[] keyArgs, Object[] valueArgs)\r\n at Microsoft.Web.Redis.RedisConnectionWrapper.TryCheckWriteLockAndGetData(Object& lockId, ISessionStateItemCollection& data, Int32& sessionTimeout)\r\n at Microsoft.Web.Redis.RedisSessionStateProvider.GetItemFromSessionStore(Boolean isWriteLockRequired, HttpContext context, String id, Boolean& locked, TimeSpan& lockAge, Object& lockId, SessionStateActions& actions)\r\n at Microsoft.Web.Redis.RedisSessionStateProvider.GetItem(HttpContext context, String id, Boolean& locked, TimeSpan& lockAge, Object& lockId, SessionStateActions& actions)\r\n at PX.Data.PXSessionStateStore.ProviderGetItem(HttpContext context, String id, Boolean& locked, TimeSpan& lockAge, Object& lockId, SessionStateActions& actionFlags)\r\n at PX.Data.PXSessionStateStore.LoadSessionItem(HttpContext context, String id, Boolean isReadonly, Object& lockId, Boolean& isNew)\r\n at PX.Data.PXSessionStateStore.WriteToSession(String id, Boolean isReadonly, Action`1 processMethod)\r\n at PX.Data.PXSessionStateStore.ProcessWithSessionContext(String id, Boolean isReadonly, Action`1 processMethod)\r\n at PX.Data.PXLongOperation.PXTaskPool.ReadSession(Object key, Boolean& abort)\r\n at PX.Data.PXLongOperation.PXTaskPool.TryGetValue(Object key, PXAsyncResult& result)\r\n at PX.Data.PXLongOperation.ClearStatus(Object key, Boolean abort)\r\n at PX.Data.PXLongOperation.ClearStatus(Object key)\r\n at PX.Data.PXGraph.Clear(PXClearOption option)\r\n at Visma.net.ERP.Api.Base.BaseRepository.d__4.MoveNext() in D:\\Data\\BuildAgent\\work\\11023e52ea15d342\\Web\\Lib\\Visma.net.ERP.Api\\Base\\BaseRepository.cs:line 77\r\n at Visma.net.ERP.Api.CA.Repository.PaymentMethodRepository.d__12.MoveNext() in D:\\Data\\BuildAgent\\work\\11023e52ea15d342\\Web\\Lib\\Visma.net.ERP.Api\\CA\\Repository\\PaymentMethodRepository.cs:line 129\r\n at Visma.net.ERP.Web.Api.Mapping.PaymentMethodMappingExtension.d__1.MoveNext() in D:\\Data\\BuildAgent\\work\\11023e52ea15d342\\Web\\Lib\\Visma.net.ERP.Web.Api\\Mapping\\PaymentMethodMappingExtension.cs:line 31\r\n at System.Linq.Buffer`1..ctor(IEnumerable`1 source)\r\n at System.Linq.Enumerable.ToArray[TSource](IEnumerable`1 source)\r\n at Visma.net.ERP.Web.Api.Mapping.PaymentMethodMappingExtension.ToDtoPaymentMethod(IPaymentMethodVo paymentMethodVo) in D:\\Data\\BuildAgent\\work\\11023e52ea15d342\\Web\\Lib\\Visma.net.ERP.Web.Api\\Mapping\\PaymentMethodMappingExtension.cs:line 16\r\n at System.Linq.Enumerable.WhereSelectEnumerableIterator`2.MoveNext()\r\n at Newtonsoft.Json.Serialization.JsonSerializerInternalWriter.SerializeList(JsonWriter writer, IEnumerable values, JsonArrayContract contract, JsonProperty member, JsonContainerContract collectionContract, JsonProperty containerProperty)\r\n at Newtonsoft.Json.Serialization.JsonSerializerInternalWriter.Serialize(JsonWriter jsonWriter, Object value, Type objectType)\r\n at Newtonsoft.Json.JsonSerializer.SerializeInternal(JsonWriter jsonWriter, Object value, Type objectType)\r\n at System.Net.Http.Formatting.BaseJsonMediaTypeFormatter.WriteToStream(Type type, Object value, Stream writeStream, Encoding effectiveEncoding)\r\n at System.Net.Http.Formatting.JsonMediaTypeFormatter.WriteToStream(Type type, Object value, Stream writeStream, Encoding effectiveEncoding)\r\n at System.Net.Http.Formatting.BaseJsonMediaTypeFormatter.WriteToStreamAsync(Type type, Object value, Stream writeStream, HttpContent content, TransportContext transportContext, CancellationToken cancellationToken)\r\n--- End of stack trace from previous location where exception was thrown ---\r\n at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()\r\n at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)\r\n at PX.Api.Compression.BaseCompressor.d__4.MoveNext()\r\n--- End of stack trace from previous location where exception was thrown ---\r\n at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()\r\n at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)\r\n at PX.Api.Compression.CompressedContent.d__4.MoveNext()\r\n--- End of stack trace from previous location where exception was thrown ---\r\n at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()\r\n at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)\r\n at System.Web.Http.WebHost.HttpControllerHandler.d__22.MoveNext()"}}
And also some 5520
Url: /controller/api/v1/carrier/
Content received:
{"ExceptionType":"IPPException","ExceptionMessage":"","ExceptionFaultCode":"5520","ExceptionMessageID":"5520_ee4a2531-eeb8-4d09-96ab-43322a86647b","ExceptionDetails":""}
Worked great for a long time, but got a bunch of 504 now
Our integrations run without any problems now
Hi,
Are you receiving errors after retry or is it only the singular call?
We are still monitoring the issue on our end.
Now 502 pops up again
Yes, working again now. Was down for almost 20 minutes ... but it is really slow sometimes. Response time from less than a second to 15 ...
Totally unavailable now, and has been like that for the last 10 minutes:
And now at 07:37
About 300 new 5520-errors logged for me the last 12 hours. Looks like it occurs in periods. Quite a few between 06:40 and 06:50. Then again around 04:00.
Different clients and endpoints
Its a lot better, but there are still quite a few 5520-errors. These are the last three that i have logged
18:13: Completed: /controller/api/v1/Inventory?availabilityLastModifiedDateTimeCondition=%3E&availabilityLastModifiedDateTime=2023-10-23 18:02:18&pageNumber=1
{"ExceptionType":"IPPException","ExceptionMessage":"","ExceptionFaultCode":"5520","ExceptionMessageID":"5520_45f4518c-0a8f-453b-9392-9a94fe84de41","ExceptionDetails":""}
18:09: Completed: /controller/api/v1/Inventory?availabilityLastModifiedDateTimeCondition=%3E&availabilityLastModifiedDateTime=2023-10-23 18:03:25&pageNumber=2
{"ExceptionType":"IPPException","ExceptionMessage":"","ExceptionFaultCode":"5520","ExceptionMessageID":"5520_c9ed6b64-c339-4e30-8f97-2241534c9675","ExceptionDetails":""}
18:07: Completed: /controller/api/v1/carrier/
{"ExceptionType":"IPPException","ExceptionMessage":"","ExceptionFaultCode":"5520","ExceptionMessageID":"5520_93028f4b-ae94-460c-b38c-1b6045af90aa","ExceptionDetails":""}
I can see that the status has changed to Operational, but that is a bit premature
From servicelog
Error: InternalServerError
Error: VismaId: 3593668b-3ed6-4c57-8eb1-466ba02181b4. Error updating supplier invoice. A connection was successfully established with the server, but then an error occurred during the pre-login handshake. (provider: SSL Provider, error: 0 - The wait operation timed out.)
IPP exception during the week end
We have identified the same problem as above this morning along with a 502-gateway error. Is this something Visma are aware of?
These seem to mainly occur when handling supplier invoices. Examples of the errors below:
{"ExceptionType":"IPPException","ExceptionMessage":"","ExceptionFaultCode":"5520","ExceptionMessageID":"5520_f468c19a-16fa-4251-ac21-928d180061dd","ExceptionDetails":""}.
{"message":"VismaId: f22461e8-9168-43b8-baf6-a387b230acb0. A connection was successfully established with the server, but then an error occurred during the pre-login handshake. (provider: SSL Provider, error: 0 - The wait operation timed out.)"}. .
Error response: <html>
<head><title>502 Bad Gateway</title></head>
<body>
<center><h1>502 Bad Gateway</h1></center>
</body>
</html>
Copyright Ā© 2022 Visma.com. All rights reserved.