Azure data factory job failing with Job Completed Batch error code - azure-data-factory

Azure Data factory pipeline activity is failing with the error "The specified job is already in a completed state.". Even re-running the activity is also failing with same error.
There are no stderr.txt and stdout.txt files generated for the tasks to gain additional insight into the issue.
"statusCode": 409,
"statusMessage": "The specified job is already in a completed state.",
"batchErrorCode": "JobCompleted",
"batchErrorMessage": "The specified job is already in a completed state."
Can anyone help us how to resolve this issue ?
Thanks in advance.

Related

Kogito Kafka messages - Message Trigger information is not complete TriggerMetaData

has someone has ever encounter this error when using Kogito (with quarkus) to send Kafka messages from a flow (bpmn2 file) throwing an "Intermediate Message" from the intermediate events?
[ERROR] com.package.xxxxxx.whenRunTestSupermarketSuccess Time elapsed: 0.006 s
<<< ERROR!
java.lang.RuntimeException:
java.lang.RuntimeException: io.quarkus.builder.BuildException: Build failure: Build failed due to errors
[error]: Build step org.kie.kogito.quarkus.common.deployment.KogitoAssetsProcessor#generateSources
threw an exception: org.kie.kogito.codegen.process.ProcessCodegenException: Error while elaborating
process id = "supermarketFlow", packageName = "com.package.xxxxxx": Message Trigger information is
not complete TriggerMetaData [name=cartCreated, type=ProduceMessage, dataType=String, modelRef=null, ownerId=1]
at org.kie.kogito.codegen.process.ProcessCodegen.internalGenerate(ProcessCodegen.java:325)
at org.kie.kogito.codegen.core.AbstractGenerator.generate(AbstractGenerator.java:69)
I can sense modelRef=null could be the issue but as far as I've seen in the official documentation, is not mention. I suppose I should give it a value on the bpmn file but I don't where.
I'm trying to send Kafka messages after running some logic, for the moment no one consuming the messages.
May be the issue is related to this thread?
BPMN 2.0, kogito, triggering signal or message from another process

Azure Devops - Release Pipeline when re-running failed tests azure devops shows failure status even if re-run succeeded

I use Specflow with SpecRunner+ I am using the Deafult.srprofile to to re-run failed tests 3 times in visual studio it shows 2passed 1 failed but the status of the test is a failure, the same goes for azure devops if a re-ran test passes the outcome of the run is a failure. The Failures are sometimes caused by locator timeouts or server timeouts not often but saw it happen few time thats why we decided to implement a re-run.
Could anyone help on this?
022-02-09T12:40:13.8607507Z Test Run Failed.
2022-02-09T12:40:13.8608607Z Total tests: 37
2022-02-09T12:40:13.8609271Z Passed: 36
2022-02-09T12:40:13.8609858Z Failed: 1
2022-02-09T12:40:13.8617476Z Total time: 7.4559 Minutes
2022-02-09T12:40:13.9226929Z ##[warning]Vstest failed with error. Check logs for failures. There might be failed tests.
2022-02-09T12:40:14.0075402Z ##[error]Error: The process 'D:\Microsoft_Visual_Studio\2019\Common7\IDE\Extensions\TestPlatform\vstest.console.exe' failed with exit code 1
2022-02-09T12:40:14.8164576Z ##[error]VsTest task failed.
But then the report states that it was retried 3 times which 2 of the retries were seccusefull but still a failure status on the azure devops run.
The behavior of the report is the correct one and sadly this can't be configured to be changed.
What you can do is to adjust how the results are reported back to Azure DevOps.
You can configure it via the VSTest element in the srProfile- File.
This example means, that at least one retry has to be passing:
<VSTest testRetryResults="Unified" passRateAbsolute="1"/>
Docs: https://docs.specflow.org/projects/specflow-runner/en/latest/Profile/VSTest.html
Be aware that we have stopped the development of the SpecFlow+ Runner. More details here: https://specflow.org/using-specflow/the-retirement-of-specflow-runner/

The slice of type 'Discovery' is 'Aborted' because of the error : System.Exception: NUnit Adapter 4.0.0.0: Test discovery complete

I tried to run Azure DevOps tests in a build pipeline.
Tests are executed on a new agent, i got the following error.
Setup Azure DevOps
##[error]The slice of type 'Discovery' is 'Aborted' because of the error : System.Exception: NUnit Adapter 4.0.0.0: Test discovery complete
Received the command : Stop
TestExecutionHost.ProcessCommand. Stop Command handled
SliceFetch Aborted. Moving to the TestHostEnd phase
Test run '1007278' is in 'Aborted' state.
##[error]Test run is aborted. Logging details of the run logs.
##[error]System.Exception: The test run was aborted, failing the task.
The Problem is that after slice process the test case filter isn't working, solution was the rename to “TestCategory”, before it was “Category“ for TestCaseFilter

Is it possible to run more than one command at the same time on Azure Data Explorer without increasing the Instance Count?

I have a scenario in which I need to export multiple CSV's from the Azure Data Explorer for that I am using the .export Command. When I try to run this request multiple times I get the following error
*TooManyRequests (429-TooManyRequests): {
"error": {
"code": "Too many requests",
"message": "Request is denied due to throttling.",
"#type": "Kusto.DataNode.Exceptions.ControlCommandThrottledException",
"#message": "The control command was aborted due to throttling. Retrying after some backoff might succeed. CommandType: 'DataExportToFile'*
Is there a way I can handle this without increasing the Instance count.
you can alter the capacity policy: https://learn.microsoft.com/en-us/azure/data-explorer/kusto/management/capacitypolicy

SAM Deployment failed Error- Waiter StackCreateComplete failed: Waiter encountered a terminal failure state

When I try to deploy package on SAM, the very first status comes in cloud formation console is ROLLBACK_IN_PROGRESS after that it gets changed to ROLLBACK_COMPLETE
I have tried deleting the stack and trying again, but every time same issue occurs.
Error in terminal looks like this-
Sourcing local options from ./SAMToolkit.devenv
SAM_PARAM_PKG environment variable not set
SAMToolkit will operate in legacy mode.
Please set SAM_PARAM_PKG in your .devenv file to run modern packaging.
Run 'sam help package' for more information
Runtime: java
Attempting to assume role from AWS Identity Broker using account 634668058279
Assumed role from AWS Identity Broker successfully.
Deploying stack sam-dev* from template: /home/***/1.0/runtime/sam/template.yml
sam-additional-artifacts-url.txt was not found, which is fine if there is no additional artifacts uploaded
Replacing BATS::SAM placeholders in template...
Uploading template build/private/tmp/sam-toolkit.yml to s3://***/sam-toolkit.yml
make_bucket failed: s3://sam-dev* An error occurred (BucketAlreadyOwnedByYou) when calling the CreateBucket operation: Your previous request to create the named bucket succeeded and you already own it.
upload: build/private/tmp/sam-toolkit.yml to s3://sam-dev*/sam-toolkit.yml
An error occurred (ValidationError) when calling the DescribeStacks operation: Stack with id sam-dev* does not exist
sam-dev* will be created.
Creating ChangeSet ChangeSet-2020-01-20T12-25-56Z
Deploying stack sam-dev*. Follow in console: https://aws-identity-broker.amazon.com/federation/634668058279/CloudFormation
ChangeSet ChangeSet-2020-01-20T12-25-56Z in sam-dev* succeeded
"StackStatus": "REVIEW_IN_PROGRESS",
sam-dev* reached REVIEW_IN_PROGRESS
Deploying stack sam-dev*. Follow in console: https://console.aws.amazon.com/cloudformation/home?region=us-west-2
Waiting for stack-create-complete
Waiter StackCreateComplete failed: Waiter encountered a terminal failure state
Command failed.
Please see the logs above.
I set SQS as event source for Lambda, but didn't provided the permissions like this
- Effect: Allow
Action:
- sqs:ReceiveMessage
- sqs:DeleteMessage
- sqs:GetQueueAttributes
Resource: "*"
in lambda policies.
I found this error in "Events" tab of "CloudFormation" service.