# Configuration

```
{
    "id": "io.infomaker.wingman",
    "name": "im-wingman",
    "style": "https://plugins.writer.infomaker.io/v1/infomaker/im-wingman/3.0.0/style.css",
    "url": "https://plugins.writer.infomaker.io/v1/infomaker/im-wingman/3.0.0/index.js",
    "mandatory": false,
    "enabled": true,
    "data": {
        "host": "https://ai-eu-west-1.saas-prod.infomaker.io",
        "contextLimit": 2,
        "encryptedKeyPhrase": "somePhrase",
        "widgets": [
            "generic",
            "headline",
            "summary"
        ],
        "widgetConfig": {
            "headline": {
                "digital": {
                    "preText": "Generate a headline",
                    "creativity": 5,
                    "headlineCount": 10,
                    "digitalHeadlineWordCount": 20,
                    "checkBoxDefault": false,
                    "providerAccessToken": "someToken",
                    "serviceProvider": "openai",
                    "modelId": "gpt-3.5-turbo"
                },
                "print": {
                    "preText": "Suggest headlines for the article provided in XML tag. If the article has a strong local connection, reflect this in the headline.",
                    "creativity": 5,
                    "headlineCount": 10,
                    "printHeadlineWordCount": 8,
                    "checkBoxDefault": false,
                    "providerAccessToken": "",
                    "serviceProvider": "Bedrock",
                    "modelId": "anthropic.claude-3-sonnet-20240229-v1:0"
                }
            },
            "summary": {
                "digital": {
                    "preText": "Generate a summary",
                    "creativity": 5,
                    "summaryCount": 5,
                    "digitalSummaryWordCount": 40,
                    "checkBoxDefault": false,
                    "providerAccessToken": "someToken",
                    "serviceProvider": "openai",
                    "modelId": "gpt-3.5-turbo"
                },
                "print": {
                    "preText": "Suggest summaries for the article provided in XML tag. Act as a news editor and your task is to suggest a summary for the article.",
                    "creativity": 5,
                    "summaryCount": 5,
                    "printSummaryWordCount": 100,
                    "checkBoxDefault": false,
                    "providerAccessToken": "",
                    "serviceProvider": "Bedrock",
                    "modelId": "anthropic.claude-v2:1"
                }
            },
            "generic": {
                "digital": {
                    "preText": "Generate a headline",
                    "creativity": 5,
                    "checkBoxDefault": false,
                    "providerAccessToken": "someToken",
                    "serviceProvider": "openai",
                    "modelId": "gpt-3.5-turbo"
                },
                "print": {
                    "preText": "You are a News Editor of a News Firm and your task is to suggest headlines for the article provided in  XML tag.\n\nPlease use professional tone while generating headlines.",
                    "creativity": 5,
                    "checkBoxDefault": false,
                    "providerAccessToken": "",
                    "serviceProvider": "Bedrock",
                    "modelId": "anthropic.claude-3-sonnet-20240229-v1:0"
                }
            }
        }
    }
}
```

### Configuration Details:

```
"data": {
        "host": "https://ai-eu-west-1.saas-prod.infomaker.io",
        "contextLimit": 2,
        "encryptedKeyPhrase": "somePhrase",
        "widgets": [
            "generic",
            "headline",
            "summary"
        ],
```

`contextLimit` : Max. limit of context tags that can be sent in the prompt.\
`encryptedKeyPhrase`: Encryption Key Phrase for sending encrypted API Key.\
`widgets` : Widgets that should be enabled in Wingman plugin\
\
**Headline Configuration Details:**

```
 "widgetConfig": {
            "headline": {
                "digital": {
                    "preText": "Generate a headline",
                    "creativity": 5,
                    "headlineCount": 10,
                    "digitalHeadlineWordCount": 20,
                    "checkBoxDefault": false,
                    "providerAccessToken": "someToken",
                    "serviceProvider": "openai",
                    "modelId": "gpt-3.5-turbo"
                },
                "print": {
                    "preText": "Suggest headlines for the article provided in XML tag. If the article has a strong local connection, reflect this in the headline.",
                    "creativity": 5,
                    "headlineCount": 10,
                    "printHeadlineWordCount": 8,
                    "checkBoxDefault": false,
                    "providerAccessToken": "",
                    "serviceProvider": "Bedrock",
                    "modelId": "anthropic.claude-3-sonnet-20240229-v1:0"
                }
            },
```

`PreText` : This is the field where the user will send the custom prompt to generate the headline.\
\&#xNAN;**\*Note:** *In case of AWS Bedrock please end your custom prompt with these words - ‘in the XML tag’*\
*Do not add any these words for ChatGPT models.*

`creativity` : This determines the temperature, between 1-5, which defines randomness in result. 1\
being least to 5 being highest.\
`headlineCount` : The total headline results to generate.\
`digitalHeadlineWordCount` : The headline result word count (approx.) to generate for digital headline\
`printHeadlineWordCount` : The headline result word count (approx.) to generate for print headline\
`checkBoxDefault: false` : By default check box should be checked or not\
`providerAccessToken` : \*Currently applicable in case of ChatGPT where if an organization want to use their own API Key for using ChatGPT service instead of Naviga's (if available)\
`serviceProvider` : Specify the AI service provider for prompt ( *openai* for ChatGPT & *Bedrock* for AWS).\
`modelID` : For a particular AI service provider the model to be used for the prompt.

**Summary Configuration Details:**

```
 "summary": {
                "digital": {
                    "preText": "Generate a summary",
                    "creativity": 5,
                    "summaryCount": 5,
                    "digitalSummaryWordCount": 40,
                    "checkBoxDefault": false,
                    "providerAccessToken": "someToken",
                    "serviceProvider": "openai",
                    "modelId": "gpt-3.5-turbo"
                },
"print": {
                    "preText": "Suggest summaries for the article provided in XML tag. Act as a news editor and your task is to suggest a summary for the article.",
                    "creativity": 5,
                    "summaryCount": 5,
                    "printSummaryWordCount": 100,
                    "checkBoxDefault": false,
                    "providerAccessToken": "",
                    "serviceProvider": "Bedrock",
                    "modelId": "anthropic.claude-v2:1"
                }
            },
```

`PreText` : This is the field where the user will send the custom prompt to generate the summary.\
\&#xNAN;**\*Note:** *In case of AWS Bedrock it is recommended to end your custom prompt with these words - ‘in the XML tag’*\
*Do not add any these words for ChatGPT models.*

`creativity` : This determines the temperature, between 1-5, which defines randomness in result. 1\
being least to 5 being highest.\
`summaryCount` : The total summary results to generate

`digitalSummaryWordCount` : The summary result word count (approx.) to generate for digital summary\
`printSummaryWordCount` : The summary result word count (approx.) to generate for print summary\
`displayCount` : Display the number of results before show more button\
`checkBoxDefault: false` : By default check box should be checked or not\
`providerAccessToken` : \*Currently applicable in case of ChatGPT where if an organization want to use their own API Key for using ChatGPT service instead of Naviga's (if available)\
`serviceProvider` : Specify the AI service provider for prompt ( *openai* for ChatGPT & *Bedrock* for AWS).\
`modelID` : For a particular AI service provider the model to be used for the prompt.

**Generic Configuration Details:**

```
"generic": {
                "digital": {
                    "preText": "Generate a headline",
                    "creativity": 5,
                    "checkBoxDefault": false,
                    "providerAccessToken": "someToken",
                    "serviceProvider": "openai",
                    "modelId": "gpt-3.5-turbo"
                },
                "print": {
                    "preText": "You are a News Editor of a News Firm and your task is to suggest headlines for the article provided in  XML tag.\n\nPlease use professional tone while generating headlines.",
                    "creativity": 5,
                    "checkBoxDefault": false,
                    "providerAccessToken": "",
                    "serviceProvider": "Bedrock",
                    "modelId": "anthropic.claude-3-sonnet-20240229-v1:0"
                }
            }
        }
    }
}
```

`PreText` : This is the field where the user will send the custom prompt to generate the generic results.\
\&#xNAN;**\*Note:** *In case of AWS Bedrock it is recommended to end your custom prompt with these words - ‘in the XML tag’*\
*Do not add any these words for ChatGPT models.*

`creativity` : This determines the temperature, between 1-5, which defines randomness in result. 1\
being least to 5 being highest.\
`checkBoxDefault : false` : By default check box should be checked or not\
`providerAccessToken` : \*Currently applicable in case of ChatGPT where if an organization want to use their own API Key for using ChatGPT service instead of Naviga's (if available)\
`serviceProvider` : Specify the AI service provider for prompt ( *openai* for ChatGPT & *Bedrock* for AWS).\
`modelID` : For a particular AI service provider the model to be used for the prompt.\ <br>
