Skip to content

Commit

Permalink
clear logs
Browse files Browse the repository at this point in the history
  • Loading branch information
xczhai committed Nov 21, 2024
1 parent ca8f26b commit c7601f3
Show file tree
Hide file tree
Showing 7 changed files with 0 additions and 1,262 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -410,11 +410,8 @@ bool ov::pass::ConvertPrecision::run_on_model(const std::shared_ptr<ov::Model>&
if (m_keep_precision_sensitive_in_fp32 && has_fp16_compression) {
pass::Manager manager(get_pass_config(), "KeepPrecisionSensitiveInFP32");
// Mark subgraphs with disable_fp16_compression to keep them in FP32
// manager.register_pass<pass::Serialize>("opt1_1.xml", "");
manager.register_pass<pass::MarkSugraphsToKeepInMixedPrecision>();
// manager.register_pass<pass::Serialize>("opt1_2.xml", "");
manager.register_pass<pass::AlignMixedFP32FP16Types>();
// manager.register_pass<pass::Serialize>("opt1_3.xml", "");
manager.run_passes(f);
}

Expand Down Expand Up @@ -494,11 +491,8 @@ bool ov::pass::ConvertPrecision::run_on_model(const std::shared_ptr<ov::Model>&
// to remove extra converts
if (m_keep_precision_sensitive_in_fp32) {
pass::Manager manager(get_pass_config(), "KeepPrecisionSensitiveInFP32:RemoveConverts");
// manager.register_pass<pass::Serialize>("opt1_4.xml", "");
manager.register_pass<pass::EnableDecompressionConvertConstantFolding>();
// manager.register_pass<pass::Serialize>("opt1_5.xml", "");
manager.register_pass<pass::ConstantFolding>();
// manager.register_pass<pass::Serialize>("opt1_6.xml", "");
manager.run_passes(f);
}

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -34,7 +34,6 @@ bool ov::pass::AlignMixedFP32FP16Types::run_on_model(const std::shared_ptr<ov::M
for (const auto& input : node->inputs()) {
const auto& incoming_output = input.get_source_output();
const auto& incoming_node = incoming_output.get_node_shared_ptr();
const auto& node_name = incoming_node->get_friendly_name();

if (fp16_compression_is_disabled(incoming_node))
continue; // we are in the middle
Expand Down
3 changes: 0 additions & 3 deletions src/plugins/intel_cpu/src/graph.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -40,7 +40,6 @@
#include "utils/node_dumper.h"
#include "utils/verbose.h"
#include "utils/precision_support.h"
#include "utils/linux_perf.hpp"

#include <oneapi/dnnl/dnnl.hpp>
#include "common/primitive_desc_iface.hpp"
Expand Down Expand Up @@ -1375,7 +1374,6 @@ void Graph::InferDynamic(SyncInferRequest* request, int numaId, UpdateStrategy&&

for (; inferCounter < stopIndx; ++inferCounter) {
auto& node = m_executableGraphNodes[inferCounter];
auto prof = LinuxPerf::Profile(node->getTypeStr());

ExecuteNodeWithCatch(node, request, numaId);
}
Expand All @@ -1396,7 +1394,6 @@ static int GetNumaNodeId(const GraphContext::CPtr& context) {
void Graph::Infer(SyncInferRequest* request) {
DEBUG_LOG("Infer graph: ", GetName(), ". Status: ", static_cast<int>(status));
const int numaId = GetNumaNodeId(m_context);
auto prof = LinuxPerf::Profile("Graph::Infer");

if (!m_pMemoryControl) {
OPENVINO_THROW("Memory control unit is not initilized in graph: ", GetName());
Expand Down
2 changes: 0 additions & 2 deletions src/plugins/intel_cpu/src/nodes/llm_mlp.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -439,8 +439,6 @@ struct LLMMLP::Executor : public LLMMLP::ExecutorBase {
int M = shape_size(ishape) / ishape[ishape.size() - 1];

auto output = m_pnode->getDstMemoryAtPort(0);
auto out_prec = output->getPrecision();
// need to cast to target precision
auto* dstC = output->getDataAs<U>();
const auto& dstStrides = output->getDescWithType<BlockedMemoryDesc>()->getStrides();
int strideC = dstStrides[dstStrides.size() - 2] * sizeof(U);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -31,7 +31,6 @@ FcConvertFusion::FcConvertFusion() {
const auto& m_a = pattern_map.at(a).get_node_shared_ptr();
const auto& m_b = pattern_map.at(b).get_node_shared_ptr();
const auto& m_fc = pattern_map.at(fc).get_node_shared_ptr();

const auto& m_convert = pattern_map.at(convert).get_node_shared_ptr();
auto output_type = m_convert->get_output_element_type(0);

Expand Down
Loading

0 comments on commit c7601f3

Please sign in to comment.