#
4. Checking Training Results
Running the train_mistral.py
script, as in the previous section, will save the resulting model in the mistral_code_generation
directory. This is a pure PyTorch model parameter file and is fully compatible with regular GPU servers, not just the MoAI Platform.
You can test the trained model using the inference_mistral.py
script located in the tutorial
directory of the GitHub repository you downloaded earlier. In this test, the prompt "Create a function that takes a list of strings as input and joins them with spaces" was used.
# tutorial/inference_mistral.py
...
input_text = """Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
Create a function to join given list of strings with space.
### Input:\n['I', 'love', 'you']
### Output:
"""
Run the code below.
~/quickstart$ python tutorial/inference_mistral.py
Upon examining the output, you can confirm that the model has appropriately generated the function as per the prompt.
Mistral: Below is an instruction that describes a task. Write a response that appropriately completes the request.
### Instruction:
Create a function to join given list of strings with space.
### Input:
['I', 'love', 'Moreh']
### Output:
def join_strings(string_list):
return ' '.join(string_list)
result = join_strings(['I', 'love', 'Moreh'])
print(result)