Skip to content

Commit 30529c1

Browse files
Investigate refactoring opportunities for batch management in Plugin and Compiler - no metadata changes
1 parent 21eb1ef commit 30529c1

File tree

5 files changed

+1
-29
lines changed

5 files changed

+1
-29
lines changed

src/plugins/intel_npu/src/backend/src/zero_infer_request.cpp

Lines changed: 1 addition & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -81,10 +81,6 @@ std::optional<size_t> determine_dynamic_batch_size(const IODescriptor& desc,
8181
return std::nullopt;
8282
}
8383

84-
if (!desc.shapeFromIRModel.has_value() || !desc.shapeFromIRModel.value().is_dynamic()) {
85-
return std::nullopt;
86-
}
87-
8884
if (batchSize.has_value()) {
8985
return batchSize.value();
9086
}
@@ -93,9 +89,7 @@ std::optional<size_t> determine_dynamic_batch_size(const IODescriptor& desc,
9389
return std::nullopt;
9490
}
9591

96-
if ((*desc.shapeFromIRModel)[intel_npu::utils::BATCH_AXIS].is_dynamic()) {
97-
return tensor->get_shape()[intel_npu::utils::BATCH_AXIS];
98-
}
92+
return tensor->get_shape()[intel_npu::utils::BATCH_AXIS];
9993

10094
return std::nullopt;
10195
}

src/plugins/intel_npu/src/common/include/intel_npu/common/igraph.hpp

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -36,8 +36,6 @@ class IGraph : public std::enable_shared_from_this<IGraph> {
3636

3737
virtual void set_argument_value(uint32_t argi, const void* argv) const = 0;
3838

39-
virtual void set_metadata(NetworkMetadata metadata) = 0;
40-
4139
virtual void initialize(const Config& config) = 0;
4240

4341
virtual ~IGraph() = default;

src/plugins/intel_npu/src/compiler_adapter/include/graph.hpp

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -35,8 +35,6 @@ class Graph : public IGraph {
3535

3636
void set_argument_value(uint32_t argi, const void* argv) const override;
3737

38-
void set_metadata(NetworkMetadata metadata) override;
39-
4038
void initialize(const Config& config) override;
4139

4240
const NetworkMetadata& get_metadata() const override;

src/plugins/intel_npu/src/compiler_adapter/src/graph.cpp

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -42,10 +42,6 @@ Graph::Graph(const std::shared_ptr<ZeGraphExtWrappers>& zeGraphExt,
4242
}
4343
}
4444

45-
void Graph::set_metadata(NetworkMetadata metadata) {
46-
_metadata = metadata;
47-
}
48-
4945
const NetworkMetadata& Graph::get_metadata() const {
5046
return _metadata;
5147
}

src/plugins/intel_npu/src/plugin/src/plugin.cpp

Lines changed: 0 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -655,18 +655,14 @@ std::shared_ptr<ov::ICompiledModel> Plugin::compile_model(const std::shared_ptr<
655655
localConfig.update({{ov::intel_npu::batch_mode.name(), strStream.str()}});
656656
}
657657

658-
bool modelDeBached = false;
659-
ov::Dimension originalBatch;
660658
if (localConfig.isAvailable(ov::intel_npu::batch_mode.name()) && modelForCompilation->is_dynamic()) {
661659
bool autoOrPluginBatch = localConfig.get<BATCH_MODE>() == ov::intel_npu::BatchMode::PLUGIN ||
662660
localConfig.get<BATCH_MODE>() == ov::intel_npu::BatchMode::AUTO;
663661
bool pluginBatchingIsSupported = validateModelBatch(modelForCompilation, _logger);
664662
if (autoOrPluginBatch && pluginBatchingIsSupported) {
665663
try {
666664
_logger.info("Attempting to handle batching on the plugin side.");
667-
originalBatch = ov::get_batch(modelForCompilation);
668665
ov::set_batch(modelForCompilation, 1);
669-
modelDeBached = true;
670666
} catch (const std::exception& ex) {
671667
_logger.info("Couldn't reshape the model. Batching will be handed by compiler.", ex.what());
672668
}
@@ -745,16 +741,6 @@ std::shared_ptr<ov::ICompiledModel> Plugin::compile_model(const std::shared_ptr<
745741
OPENVINO_THROW("NPU plugin: got an unexpected exception from compiler");
746742
}
747743

748-
if (modelDeBached) {
749-
auto metadata = graph->get_metadata();
750-
for (auto& in : metadata.inputs) {
751-
if (in.shapeFromIRModel.has_value() && originalBatch.get_max_length() != 1) {
752-
in.shapeFromIRModel.value()[intel_npu::utils::BATCH_AXIS] = originalBatch;
753-
}
754-
}
755-
graph->set_metadata(metadata);
756-
}
757-
758744
std::shared_ptr<ov::ICompiledModel> compiledModel;
759745
try {
760746
compiledModel = std::make_shared<CompiledModel>(model, shared_from_this(), device, graph, localConfig);

0 commit comments

Comments
 (0)