Remove flowlauncher from imports
Browse files
run.py
CHANGED
|
@@ -5,7 +5,6 @@ import os
|
|
| 5 |
import hydra
|
| 6 |
|
| 7 |
import aiflows
|
| 8 |
-
from aiflows.flow_launchers import FlowLauncher
|
| 9 |
from aiflows.backends.api_info import ApiInfo
|
| 10 |
from aiflows.utils.general_helpers import read_yaml_file, quick_load_api_keys
|
| 11 |
|
|
@@ -90,10 +89,10 @@ if __name__ == "__main__":
|
|
| 90 |
)
|
| 91 |
|
| 92 |
#option2: use the proxy_flow
|
| 93 |
-
#input_message = proxy_flow.
|
| 94 |
|
| 95 |
#7. ~~~ Run inference ~~~
|
| 96 |
-
future = proxy_flow.
|
| 97 |
|
| 98 |
#uncomment this line if you would like to get the full message back
|
| 99 |
#reply_message = future.get_message()
|
|
|
|
| 5 |
import hydra
|
| 6 |
|
| 7 |
import aiflows
|
|
|
|
| 8 |
from aiflows.backends.api_info import ApiInfo
|
| 9 |
from aiflows.utils.general_helpers import read_yaml_file, quick_load_api_keys
|
| 10 |
|
|
|
|
| 89 |
)
|
| 90 |
|
| 91 |
#option2: use the proxy_flow
|
| 92 |
+
#input_message = proxy_flow.package_input_message(data = data)
|
| 93 |
|
| 94 |
#7. ~~~ Run inference ~~~
|
| 95 |
+
future = proxy_flow.get_reply_future(input_message)
|
| 96 |
|
| 97 |
#uncomment this line if you would like to get the full message back
|
| 98 |
#reply_message = future.get_message()
|